123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548 |
- 'use strict';
- const MONGODB_ERROR_CODES = require('../error_codes').MONGODB_ERROR_CODES;
- const core = require('../core');
- const crypto = require('crypto');
- const stream = require('stream');
- const util = require('util');
- const Buffer = require('safe-buffer').Buffer;
- const deprecateOptions = require('../utils').deprecateOptions;
- /**
- * A writable stream that enables you to write buffers to GridFS.
- *
- * Do not instantiate this class directly. Use `openUploadStream()` instead.
- *
- * @class
- * @extends external:Writable
- * @param {GridFSBucket} bucket Handle for this stream's corresponding bucket
- * @param {string} filename The value of the 'filename' key in the files doc
- * @param {object} [options] Optional settings.
- * @param {string|number|object} [options.id] Custom file id for the GridFS file.
- * @param {number} [options.chunkSizeBytes] The chunk size to use, in bytes
- * @param {(number|string)} [options.w] **Deprecated** The write concern. Use writeConcern instead.
- * @param {number} [options.wtimeout] **Deprecated** The write concern timeout. Use writeConcern instead.
- * @param {boolean} [options.j=false] **Deprecated** Specify a journal write concern. Use writeConcern instead.
- * @param {object|WriteConcern} [options.writeConcern] Specify write concern settings.
- * @param {boolean} [options.disableMD5=false] If true, disables adding an md5 field to file data
- * @fires GridFSBucketWriteStream#error
- * @fires GridFSBucketWriteStream#finish
- */
- const GridFSBucketWriteStream = deprecateOptions(
- {
- name: 'GridFSBucketWriteStream',
- deprecatedOptions: ['disableMD5'],
- optionsIndex: 2
- },
- function(bucket, filename, options) {
- options = options || {};
- stream.Writable.call(this, options);
- this.bucket = bucket;
- this.chunks = bucket.s._chunksCollection;
- this.filename = filename;
- this.files = bucket.s._filesCollection;
- this.options = options;
- // Signals the write is all done
- this.done = false;
- this.id = options.id ? options.id : core.BSON.ObjectId();
- this.chunkSizeBytes = this.options.chunkSizeBytes;
- this.bufToStore = Buffer.alloc(this.chunkSizeBytes);
- this.length = 0;
- this.md5 = !options.disableMD5 && crypto.createHash('md5');
- this.n = 0;
- this.pos = 0;
- this.state = {
- streamEnd: false,
- outstandingRequests: 0,
- errored: false,
- aborted: false,
- promiseLibrary: this.bucket.s.promiseLibrary
- };
- if (!this.bucket.s.calledOpenUploadStream) {
- this.bucket.s.calledOpenUploadStream = true;
- var _this = this;
- checkIndexes(this, function() {
- _this.bucket.s.checkedIndexes = true;
- _this.bucket.emit('index');
- });
- }
- }
- );
- util.inherits(GridFSBucketWriteStream, stream.Writable);
- /**
- * An error occurred
- *
- * @event GridFSBucketWriteStream#error
- * @type {Error}
- */
- /**
- * `end()` was called and the write stream successfully wrote the file
- * metadata and all the chunks to MongoDB.
- *
- * @event GridFSBucketWriteStream#finish
- * @type {object}
- */
- /**
- * Write a buffer to the stream.
- *
- * @method
- * @param {Buffer} chunk Buffer to write
- * @param {String} encoding Optional encoding for the buffer
- * @param {GridFSBucket~errorCallback} callback Function to call when the chunk was added to the buffer, or if the entire chunk was persisted to MongoDB if this chunk caused a flush.
- * @return {Boolean} False if this write required flushing a chunk to MongoDB. True otherwise.
- */
- GridFSBucketWriteStream.prototype.write = function(chunk, encoding, callback) {
- var _this = this;
- return waitForIndexes(this, function() {
- return doWrite(_this, chunk, encoding, callback);
- });
- };
- /**
- * Places this write stream into an aborted state (all future writes fail)
- * and deletes all chunks that have already been written.
- *
- * @method
- * @param {GridFSBucket~errorCallback} callback called when chunks are successfully removed or error occurred
- * @return {Promise} if no callback specified
- */
- GridFSBucketWriteStream.prototype.abort = function(callback) {
- if (this.state.streamEnd) {
- var error = new Error('Cannot abort a stream that has already completed');
- if (typeof callback === 'function') {
- return callback(error);
- }
- return this.state.promiseLibrary.reject(error);
- }
- if (this.state.aborted) {
- error = new Error('Cannot call abort() on a stream twice');
- if (typeof callback === 'function') {
- return callback(error);
- }
- return this.state.promiseLibrary.reject(error);
- }
- this.state.aborted = true;
- this.chunks.deleteMany({ files_id: this.id }, function(error) {
- if (typeof callback === 'function') callback(error);
- });
- };
- /**
- * Tells the stream that no more data will be coming in. The stream will
- * persist the remaining data to MongoDB, write the files document, and
- * then emit a 'finish' event.
- *
- * @method
- * @param {Buffer} chunk Buffer to write
- * @param {String} encoding Optional encoding for the buffer
- * @param {GridFSBucket~errorCallback} callback Function to call when all files and chunks have been persisted to MongoDB
- */
- GridFSBucketWriteStream.prototype.end = function(chunk, encoding, callback) {
- var _this = this;
- if (typeof chunk === 'function') {
- (callback = chunk), (chunk = null), (encoding = null);
- } else if (typeof encoding === 'function') {
- (callback = encoding), (encoding = null);
- }
- if (checkAborted(this, callback)) {
- return;
- }
- this.state.streamEnd = true;
- if (callback) {
- this.once('finish', function(result) {
- callback(null, result);
- });
- }
- if (!chunk) {
- waitForIndexes(this, function() {
- writeRemnant(_this);
- });
- return;
- }
- this.write(chunk, encoding, function() {
- writeRemnant(_this);
- });
- };
- /**
- * @ignore
- */
- function __handleError(_this, error, callback) {
- if (_this.state.errored) {
- return;
- }
- _this.state.errored = true;
- if (callback) {
- return callback(error);
- }
- _this.emit('error', error);
- }
- /**
- * @ignore
- */
- function createChunkDoc(filesId, n, data) {
- return {
- _id: core.BSON.ObjectId(),
- files_id: filesId,
- n: n,
- data: data
- };
- }
- /**
- * @ignore
- */
- function checkChunksIndex(_this, callback) {
- _this.chunks.listIndexes().toArray(function(error, indexes) {
- if (error) {
- // Collection doesn't exist so create index
- if (error.code === MONGODB_ERROR_CODES.NamespaceNotFound) {
- var index = { files_id: 1, n: 1 };
- _this.chunks.createIndex(index, { background: false, unique: true }, function(error) {
- if (error) {
- return callback(error);
- }
- callback();
- });
- return;
- }
- return callback(error);
- }
- var hasChunksIndex = false;
- indexes.forEach(function(index) {
- if (index.key) {
- var keys = Object.keys(index.key);
- if (keys.length === 2 && index.key.files_id === 1 && index.key.n === 1) {
- hasChunksIndex = true;
- }
- }
- });
- if (hasChunksIndex) {
- callback();
- } else {
- index = { files_id: 1, n: 1 };
- var indexOptions = getWriteOptions(_this);
- indexOptions.background = false;
- indexOptions.unique = true;
- _this.chunks.createIndex(index, indexOptions, function(error) {
- if (error) {
- return callback(error);
- }
- callback();
- });
- }
- });
- }
- /**
- * @ignore
- */
- function checkDone(_this, callback) {
- if (_this.done) return true;
- if (_this.state.streamEnd && _this.state.outstandingRequests === 0 && !_this.state.errored) {
- // Set done so we dont' trigger duplicate createFilesDoc
- _this.done = true;
- // Create a new files doc
- var filesDoc = createFilesDoc(
- _this.id,
- _this.length,
- _this.chunkSizeBytes,
- _this.md5 && _this.md5.digest('hex'),
- _this.filename,
- _this.options.contentType,
- _this.options.aliases,
- _this.options.metadata
- );
- if (checkAborted(_this, callback)) {
- return false;
- }
- _this.files.insertOne(filesDoc, getWriteOptions(_this), function(error) {
- if (error) {
- return __handleError(_this, error, callback);
- }
- _this.emit('finish', filesDoc);
- _this.emit('close');
- });
- return true;
- }
- return false;
- }
- /**
- * @ignore
- */
- function checkIndexes(_this, callback) {
- _this.files.findOne({}, { _id: 1 }, function(error, doc) {
- if (error) {
- return callback(error);
- }
- if (doc) {
- return callback();
- }
- _this.files.listIndexes().toArray(function(error, indexes) {
- if (error) {
- // Collection doesn't exist so create index
- if (error.code === MONGODB_ERROR_CODES.NamespaceNotFound) {
- var index = { filename: 1, uploadDate: 1 };
- _this.files.createIndex(index, { background: false }, function(error) {
- if (error) {
- return callback(error);
- }
- checkChunksIndex(_this, callback);
- });
- return;
- }
- return callback(error);
- }
- var hasFileIndex = false;
- indexes.forEach(function(index) {
- var keys = Object.keys(index.key);
- if (keys.length === 2 && index.key.filename === 1 && index.key.uploadDate === 1) {
- hasFileIndex = true;
- }
- });
- if (hasFileIndex) {
- checkChunksIndex(_this, callback);
- } else {
- index = { filename: 1, uploadDate: 1 };
- var indexOptions = getWriteOptions(_this);
- indexOptions.background = false;
- _this.files.createIndex(index, indexOptions, function(error) {
- if (error) {
- return callback(error);
- }
- checkChunksIndex(_this, callback);
- });
- }
- });
- });
- }
- /**
- * @ignore
- */
- function createFilesDoc(_id, length, chunkSize, md5, filename, contentType, aliases, metadata) {
- var ret = {
- _id: _id,
- length: length,
- chunkSize: chunkSize,
- uploadDate: new Date(),
- filename: filename
- };
- if (md5) {
- ret.md5 = md5;
- }
- if (contentType) {
- ret.contentType = contentType;
- }
- if (aliases) {
- ret.aliases = aliases;
- }
- if (metadata) {
- ret.metadata = metadata;
- }
- return ret;
- }
- /**
- * @ignore
- */
- function doWrite(_this, chunk, encoding, callback) {
- if (checkAborted(_this, callback)) {
- return false;
- }
- var inputBuf = Buffer.isBuffer(chunk) ? chunk : Buffer.from(chunk, encoding);
- _this.length += inputBuf.length;
- // Input is small enough to fit in our buffer
- if (_this.pos + inputBuf.length < _this.chunkSizeBytes) {
- inputBuf.copy(_this.bufToStore, _this.pos);
- _this.pos += inputBuf.length;
- callback && callback();
- // Note that we reverse the typical semantics of write's return value
- // to be compatible with node's `.pipe()` function.
- // True means client can keep writing.
- return true;
- }
- // Otherwise, buffer is too big for current chunk, so we need to flush
- // to MongoDB.
- var inputBufRemaining = inputBuf.length;
- var spaceRemaining = _this.chunkSizeBytes - _this.pos;
- var numToCopy = Math.min(spaceRemaining, inputBuf.length);
- var outstandingRequests = 0;
- while (inputBufRemaining > 0) {
- var inputBufPos = inputBuf.length - inputBufRemaining;
- inputBuf.copy(_this.bufToStore, _this.pos, inputBufPos, inputBufPos + numToCopy);
- _this.pos += numToCopy;
- spaceRemaining -= numToCopy;
- if (spaceRemaining === 0) {
- if (_this.md5) {
- _this.md5.update(_this.bufToStore);
- }
- var doc = createChunkDoc(_this.id, _this.n, Buffer.from(_this.bufToStore));
- ++_this.state.outstandingRequests;
- ++outstandingRequests;
- if (checkAborted(_this, callback)) {
- return false;
- }
- _this.chunks.insertOne(doc, getWriteOptions(_this), function(error) {
- if (error) {
- return __handleError(_this, error);
- }
- --_this.state.outstandingRequests;
- --outstandingRequests;
- if (!outstandingRequests) {
- _this.emit('drain', doc);
- callback && callback();
- checkDone(_this);
- }
- });
- spaceRemaining = _this.chunkSizeBytes;
- _this.pos = 0;
- ++_this.n;
- }
- inputBufRemaining -= numToCopy;
- numToCopy = Math.min(spaceRemaining, inputBufRemaining);
- }
- // Note that we reverse the typical semantics of write's return value
- // to be compatible with node's `.pipe()` function.
- // False means the client should wait for the 'drain' event.
- return false;
- }
- /**
- * @ignore
- */
- function getWriteOptions(_this) {
- var obj = {};
- if (_this.options.writeConcern) {
- obj.w = _this.options.writeConcern.w;
- obj.wtimeout = _this.options.writeConcern.wtimeout;
- obj.j = _this.options.writeConcern.j;
- }
- return obj;
- }
- /**
- * @ignore
- */
- function waitForIndexes(_this, callback) {
- if (_this.bucket.s.checkedIndexes) {
- return callback(false);
- }
- _this.bucket.once('index', function() {
- callback(true);
- });
- return true;
- }
- /**
- * @ignore
- */
- function writeRemnant(_this, callback) {
- // Buffer is empty, so don't bother to insert
- if (_this.pos === 0) {
- return checkDone(_this, callback);
- }
- ++_this.state.outstandingRequests;
- // Create a new buffer to make sure the buffer isn't bigger than it needs
- // to be.
- var remnant = Buffer.alloc(_this.pos);
- _this.bufToStore.copy(remnant, 0, 0, _this.pos);
- if (_this.md5) {
- _this.md5.update(remnant);
- }
- var doc = createChunkDoc(_this.id, _this.n, remnant);
- // If the stream was aborted, do not write remnant
- if (checkAborted(_this, callback)) {
- return false;
- }
- _this.chunks.insertOne(doc, getWriteOptions(_this), function(error) {
- if (error) {
- return __handleError(_this, error);
- }
- --_this.state.outstandingRequests;
- checkDone(_this);
- });
- }
- /**
- * @ignore
- */
- function checkAborted(_this, callback) {
- if (_this.state.aborted) {
- if (typeof callback === 'function') {
- callback(new Error('this stream has been aborted'));
- }
- return true;
- }
- return false;
- }
- module.exports = GridFSBucketWriteStream;
|