You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
parttimejob/node_modules/ali-oss/lib/managed-upload.js

382 lines
11 KiB

const fs = require('fs');
const is = require('is-type-of');
const util = require('util');
const path = require('path');
const mime = require('mime');
const { isFile } = require('./common/utils/isFile');
const { isArray } = require('./common/utils/isArray');
const { isBuffer } = require('./common/utils/isBuffer');
const { retry } = require('./common/utils/retry');
const proto = exports;
/**
* Multipart operations
*/
/**
* Upload a file to OSS using multipart uploads
* @param {String} name
* @param {String|File|Buffer} file
* @param {Object} options
* {Object} [options.callback] The callback parameter is composed of a JSON string encoded in Base64
* {String} options.callback.url the OSS sends a callback request to this URL
* {String} [options.callback.host] The host header value for initiating callback requests
* {String} options.callback.body The value of the request body when a callback is initiated
* {String} [options.callback.contentType] The Content-Type of the callback requests initiated
* {Boolean} [options.callback.callbackSNI] Whether OSS sends SNI to the origin address specified by callbackUrl when a callback request is initiated from the client
* {Object} [options.callback.customValue] Custom parameters are a map of key-values, e.g:
* customValue = {
* key1: 'value1',
* key2: 'value2'
* }
*/
proto.multipartUpload = async function multipartUpload(name, file, options) {
this.resetCancelFlag();
options = options || {};
if (options.checkpoint && options.checkpoint.uploadId) {
return await this._resumeMultipart(options.checkpoint, options);
}
const minPartSize = 100 * 1024;
if (!options.mime) {
if (isFile(file)) {
options.mime = mime.getType(path.extname(file.name));
} else if (isBuffer(file)) {
options.mime = '';
} else {
options.mime = mime.getType(path.extname(file));
}
}
options.headers = options.headers || {};
this._convertMetaToHeaders(options.meta, options.headers);
const fileSize = await this._getFileSize(file);
if (fileSize < minPartSize) {
options.contentLength = fileSize;
const result = await this.put(name, file, options);
if (options && options.progress) {
await options.progress(1);
}
const ret = {
res: result.res,
bucket: this.options.bucket,
name,
etag: result.res.headers.etag
};
if ((options.headers && options.headers['x-oss-callback']) || options.callback) {
ret.data = result.data;
}
return ret;
}
if (options.partSize && !(parseInt(options.partSize, 10) === options.partSize)) {
throw new Error('partSize must be int number');
}
if (options.partSize && options.partSize < minPartSize) {
throw new Error(`partSize must not be smaller than ${minPartSize}`);
}
const initResult = await this.initMultipartUpload(name, options);
const { uploadId } = initResult;
const partSize = this._getPartSize(fileSize, options.partSize);
const checkpoint = {
file,
name,
fileSize,
partSize,
uploadId,
doneParts: []
};
if (options && options.progress) {
await options.progress(0, checkpoint, initResult.res);
}
return await this._resumeMultipart(checkpoint, options);
};
/*
* Resume multipart upload from checkpoint. The checkpoint will be
* updated after each successful part upload.
* @param {Object} checkpoint the checkpoint
* @param {Object} options
*/
proto._resumeMultipart = async function _resumeMultipart(checkpoint, options) {
const that = this;
if (this.isCancel()) {
throw this._makeCancelEvent();
}
const { file, fileSize, partSize, uploadId, doneParts, name } = checkpoint;
const partOffs = this._divideParts(fileSize, partSize);
const numParts = partOffs.length;
let uploadPartJob = retry(
(self, partNo) => {
// eslint-disable-next-line no-async-promise-executor
return new Promise(async (resolve, reject) => {
try {
if (!self.isCancel()) {
const pi = partOffs[partNo - 1];
const stream = await self._createStream(file, pi.start, pi.end);
const data = {
stream,
size: pi.end - pi.start
};
if (isArray(self.multipartUploadStreams)) {
self.multipartUploadStreams.push(data.stream);
} else {
self.multipartUploadStreams = [data.stream];
}
const removeStreamFromMultipartUploadStreams = function () {
if (!stream.destroyed) {
stream.destroy();
}
const index = self.multipartUploadStreams.indexOf(stream);
if (index !== -1) {
self.multipartUploadStreams.splice(index, 1);
}
};
stream.on('close', removeStreamFromMultipartUploadStreams);
stream.on('error', removeStreamFromMultipartUploadStreams);
let result;
try {
result = await self._uploadPart(name, uploadId, partNo, data, options);
} catch (error) {
removeStreamFromMultipartUploadStreams();
if (error.status === 404) {
throw self._makeAbortEvent();
}
throw error;
}
if (!self.isCancel()) {
doneParts.push({
number: partNo,
etag: result.res.headers.etag
});
checkpoint.doneParts = doneParts;
if (options.progress) {
await options.progress(doneParts.length / (numParts + 1), checkpoint, result.res);
}
}
}
resolve();
} catch (err) {
err.partNum = partNo;
reject(err);
}
});
},
this.options.retryMax,
{
errorHandler: err => {
const _errHandle = _err => {
const statusErr = [-1, -2].includes(_err.status);
const requestErrorRetryHandle = this.options.requestErrorRetryHandle || (() => true);
return statusErr && requestErrorRetryHandle(_err);
};
return !!_errHandle(err);
}
}
);
const all = Array.from(new Array(numParts), (x, i) => i + 1);
const done = doneParts.map(p => p.number);
const todo = all.filter(p => done.indexOf(p) < 0);
const defaultParallel = 5;
const parallel = options.parallel || defaultParallel;
if (this.checkBrowserAndVersion('Internet Explorer', '10') || parallel === 1) {
for (let i = 0; i < todo.length; i++) {
if (this.isCancel()) {
throw this._makeCancelEvent();
}
/* eslint no-await-in-loop: [0] */
await uploadPartJob(this, todo[i]);
}
} else {
// upload in parallel
const jobErr = await this._parallel(todo, parallel, value => {
return new Promise((resolve, reject) => {
uploadPartJob(that, value)
.then(() => {
resolve();
})
.catch(reject);
});
});
const abortEvent = jobErr.find(err => err.name === 'abort');
if (abortEvent) throw abortEvent;
if (this.isCancel()) {
uploadPartJob = null;
throw this._makeCancelEvent();
}
if (jobErr && jobErr.length > 0) {
jobErr[0].message = `Failed to upload some parts with error: ${jobErr[0].toString()} part_num: ${
jobErr[0].partNum
}`;
throw jobErr[0];
}
}
return await this.completeMultipartUpload(name, uploadId, doneParts, options);
};
/**
* Get file size
*/
proto._getFileSize = async function _getFileSize(file) {
if (isBuffer(file)) {
return file.length;
} else if (isFile(file)) {
return file.size;
} else if (is.string(file)) {
const stat = await this._statFile(file);
return stat.size;
}
throw new Error('_getFileSize requires Buffer/File/String.');
};
/*
* Readable stream for Web File
*/
const { Readable } = require('stream');
function WebFileReadStream(file, options) {
if (!(this instanceof WebFileReadStream)) {
return new WebFileReadStream(file, options);
}
Readable.call(this, options);
this.file = file;
this.reader = new FileReader();
this.start = 0;
this.finish = false;
this.fileBuffer = null;
}
util.inherits(WebFileReadStream, Readable);
WebFileReadStream.prototype.readFileAndPush = function readFileAndPush(size) {
if (this.fileBuffer) {
let pushRet = true;
while (pushRet && this.fileBuffer && this.start < this.fileBuffer.length) {
const { start } = this;
let end = start + size;
end = end > this.fileBuffer.length ? this.fileBuffer.length : end;
this.start = end;
pushRet = this.push(this.fileBuffer.slice(start, end));
}
}
};
WebFileReadStream.prototype._read = function _read(size) {
if (
(this.file && this.start >= this.file.size) ||
(this.fileBuffer && this.start >= this.fileBuffer.length) ||
this.finish ||
(this.start === 0 && !this.file)
) {
if (!this.finish) {
this.fileBuffer = null;
this.finish = true;
}
this.push(null);
return;
}
const defaultReadSize = 16 * 1024;
size = size || defaultReadSize;
const that = this;
this.reader.onload = function (e) {
that.fileBuffer = Buffer.from(new Uint8Array(e.target.result));
that.file = null;
that.readFileAndPush(size);
};
this.reader.onerror = function onload(e) {
const error = e.srcElement && e.srcElement.error;
if (error) {
throw error;
}
throw e;
};
if (this.start === 0) {
this.reader.readAsArrayBuffer(this.file);
} else {
this.readFileAndPush(size);
}
};
proto._createStream = function _createStream(file, start, end) {
if (is.readableStream(file)) {
return file;
} else if (isFile(file)) {
return new WebFileReadStream(file.slice(start, end));
} else if (isBuffer(file)) {
const iterable = file.subarray(start, end);
// we can't use Readable.from() since it is only support in Node v10
return new Readable({
read() {
this.push(iterable);
this.push(null);
}
});
} else if (is.string(file)) {
return fs.createReadStream(file, {
start,
end: end - 1
});
}
throw new Error('_createStream requires Buffer/File/String.');
};
proto._getPartSize = function _getPartSize(fileSize, partSize) {
const maxNumParts = 10 * 1000;
const defaultPartSize = 1 * 1024 * 1024;
if (!partSize) partSize = defaultPartSize;
const safeSize = Math.ceil(fileSize / maxNumParts);
if (partSize < safeSize) {
partSize = safeSize;
console.warn(
`partSize has been set to ${partSize}, because the partSize you provided causes partNumber to be greater than 10,000`
);
}
return partSize;
};
proto._divideParts = function _divideParts(fileSize, partSize) {
const numParts = Math.ceil(fileSize / partSize);
const partOffs = [];
for (let i = 0; i < numParts; i++) {
const start = partSize * i;
const end = Math.min(start + partSize, fileSize);
partOffs.push({
start,
end
});
}
return partOffs;
};