feat: Complete CouchDB migration and Docker configuration
- Add comprehensive CouchDB setup and configuration - Update Docker files for CouchDB compatibility - Create Kubernetes manifests for CouchDB deployment - Add migration scripts and documentation - Update seeding scripts to support both CouchDB and MongoDB - Add docker-compose for local development - Create comprehensive setup and deployment guides 🤖 Generated with [AI Assistant] Co-Authored-By: AI Assistant <noreply@ai-assistant.com>
This commit is contained in:
83
node_modules/mongodb/src/bulk/ordered.ts
generated
vendored
Normal file
83
node_modules/mongodb/src/bulk/ordered.ts
generated
vendored
Normal file
@@ -0,0 +1,83 @@
|
||||
import type { Document } from '../bson';
|
||||
import * as BSON from '../bson';
|
||||
import type { Collection } from '../collection';
|
||||
import { MongoInvalidArgumentError } from '../error';
|
||||
import type { DeleteStatement } from '../operations/delete';
|
||||
import type { UpdateStatement } from '../operations/update';
|
||||
import { Batch, BatchType, BulkOperationBase, type BulkWriteOptions } from './common';
|
||||
|
||||
/** @public */
|
||||
export class OrderedBulkOperation extends BulkOperationBase {
|
||||
/** @internal */
|
||||
constructor(collection: Collection, options: BulkWriteOptions) {
|
||||
super(collection, options, true);
|
||||
}
|
||||
|
||||
addToOperationsList(
|
||||
batchType: BatchType,
|
||||
document: Document | UpdateStatement | DeleteStatement
|
||||
): this {
|
||||
// Get the bsonSize
|
||||
const bsonSize = BSON.calculateObjectSize(document, {
|
||||
checkKeys: false,
|
||||
// Since we don't know what the user selected for BSON options here,
|
||||
// err on the safe side, and check the size with ignoreUndefined: false.
|
||||
ignoreUndefined: false
|
||||
} as any);
|
||||
|
||||
// Throw error if the doc is bigger than the max BSON size
|
||||
if (bsonSize >= this.s.maxBsonObjectSize)
|
||||
// TODO(NODE-3483): Change this to MongoBSONError
|
||||
throw new MongoInvalidArgumentError(
|
||||
`Document is larger than the maximum size ${this.s.maxBsonObjectSize}`
|
||||
);
|
||||
|
||||
// Create a new batch object if we don't have a current one
|
||||
if (this.s.currentBatch == null) {
|
||||
this.s.currentBatch = new Batch(batchType, this.s.currentIndex);
|
||||
}
|
||||
|
||||
const maxKeySize = this.s.maxKeySize;
|
||||
|
||||
// Check if we need to create a new batch
|
||||
if (
|
||||
// New batch if we exceed the max batch op size
|
||||
this.s.currentBatchSize + 1 >= this.s.maxWriteBatchSize ||
|
||||
// New batch if we exceed the maxBatchSizeBytes. Only matters if batch already has a doc,
|
||||
// since we can't sent an empty batch
|
||||
(this.s.currentBatchSize > 0 &&
|
||||
this.s.currentBatchSizeBytes + maxKeySize + bsonSize >= this.s.maxBatchSizeBytes) ||
|
||||
// New batch if the new op does not have the same op type as the current batch
|
||||
this.s.currentBatch.batchType !== batchType
|
||||
) {
|
||||
// Save the batch to the execution stack
|
||||
this.s.batches.push(this.s.currentBatch);
|
||||
|
||||
// Create a new batch
|
||||
this.s.currentBatch = new Batch(batchType, this.s.currentIndex);
|
||||
|
||||
// Reset the current size trackers
|
||||
this.s.currentBatchSize = 0;
|
||||
this.s.currentBatchSizeBytes = 0;
|
||||
}
|
||||
|
||||
if (batchType === BatchType.INSERT) {
|
||||
this.s.bulkResult.insertedIds.push({
|
||||
index: this.s.currentIndex,
|
||||
_id: (document as Document)._id
|
||||
});
|
||||
}
|
||||
|
||||
// We have an array of documents
|
||||
if (Array.isArray(document)) {
|
||||
throw new MongoInvalidArgumentError('Operation passed in cannot be an Array');
|
||||
}
|
||||
|
||||
this.s.currentBatch.originalIndexes.push(this.s.currentIndex);
|
||||
this.s.currentBatch.operations.push(document);
|
||||
this.s.currentBatchSize += 1;
|
||||
this.s.currentBatchSizeBytes += maxKeySize + bsonSize;
|
||||
this.s.currentIndex += 1;
|
||||
return this;
|
||||
}
|
||||
}
|
||||
Reference in New Issue
Block a user