'use strict'; const common = require('./common'); const BulkOperationBase = common.BulkOperationBase; const Batch = common.Batch; const bson = common.bson; const utils = require('../utils'); const toError = utils.toError; /** * Add to internal list of Operations * * @param {UnorderedBulkOperation} bulkOperation * @param {number} docType number indicating the document type * @param {object} document * @return {UnorderedBulkOperation} */ function addToOperationsList(bulkOperation, docType, document) { // Get the bsonSize const bsonSize = bson.calculateObjectSize(document, { checkKeys: false, // Since we don't know what the user selected for BSON options here, // err on the safe side, and check the size with ignoreUndefined: false. ignoreUndefined: false }); // Throw error if the doc is bigger than the max BSON size if (bsonSize >= bulkOperation.s.maxBatchSizeBytes) throw toError('document is larger than the maximum size ' + bulkOperation.s.maxBatchSizeBytes); // Holds the current batch bulkOperation.s.currentBatch = null; // Get the right type of batch if (docType === common.INSERT) { bulkOperation.s.currentBatch = bulkOperation.s.currentInsertBatch; } else if (docType === common.UPDATE) { bulkOperation.s.currentBatch = bulkOperation.s.currentUpdateBatch; } else if (docType === common.REMOVE) { bulkOperation.s.currentBatch = bulkOperation.s.currentRemoveBatch; } const maxKeySize = bulkOperation.s.maxKeySize; // Create a new batch object if we don't have a current one if (bulkOperation.s.currentBatch == null) bulkOperation.s.currentBatch = new Batch(docType, bulkOperation.s.currentIndex); // Check if we need to create a new batch if ( bulkOperation.s.currentBatch.size + 1 >= bulkOperation.s.maxWriteBatchSize || bulkOperation.s.currentBatch.sizeBytes + maxKeySize + bsonSize >= bulkOperation.s.maxBatchSizeBytes || bulkOperation.s.currentBatch.batchType !== docType ) { // Save the batch to the execution stack bulkOperation.s.batches.push(bulkOperation.s.currentBatch); // Create a new batch bulkOperation.s.currentBatch = new Batch(docType, bulkOperation.s.currentIndex); } // We have an array of documents if (Array.isArray(document)) { throw toError('operation passed in cannot be an Array'); } bulkOperation.s.currentBatch.operations.push(document); bulkOperation.s.currentBatch.originalIndexes.push(bulkOperation.s.currentIndex); bulkOperation.s.currentIndex = bulkOperation.s.currentIndex + 1; // Save back the current Batch to the right type if (docType === common.INSERT) { bulkOperation.s.currentInsertBatch = bulkOperation.s.currentBatch; bulkOperation.s.bulkResult.insertedIds.push({ index: bulkOperation.s.bulkResult.insertedIds.length, _id: document._id }); } else if (docType === common.UPDATE) { bulkOperation.s.currentUpdateBatch = bulkOperation.s.currentBatch; } else if (docType === common.REMOVE) { bulkOperation.s.currentRemoveBatch = bulkOperation.s.currentBatch; } // Update current batch size bulkOperation.s.currentBatch.size += 1; bulkOperation.s.currentBatch.sizeBytes += maxKeySize + bsonSize; // Return bulkOperation return bulkOperation; } /** * Create a new UnorderedBulkOperation instance (INTERNAL TYPE, do not instantiate directly) * @class * @extends BulkOperationBase * @property {number} length Get the number of operations in the bulk. * @return {UnorderedBulkOperation} a UnorderedBulkOperation instance. */ class UnorderedBulkOperation extends BulkOperationBase { constructor(topology, collection, options) { options = options || {}; options = Object.assign(options, { addToOperationsList }); super(topology, collection, options, false); } } /** * Returns an unordered batch object * @ignore */ function initializeUnorderedBulkOp(topology, collection, options) { return new UnorderedBulkOperation(topology, collection, options); } initializeUnorderedBulkOp.UnorderedBulkOperation = UnorderedBulkOperation; module.exports = initializeUnorderedBulkOp; module.exports.Bulk = UnorderedBulkOperation;