mirror of
				https://github.com/actions/upload-artifact.git
				synced 2025-11-03 23:50:56 +00:00 
			
		
		
		
	bump @actions/artifact to v2.1.9
This commit is contained in:
		
					parent
					
						
							
								0b2256b8c0
							
						
					
				
			
			
				commit
				
					
						e445c64bc2
					
				
			
		
					 4 changed files with 67 additions and 71 deletions
				
			
		
							
								
								
									
										58
									
								
								dist/merge/index.js
									
										
									
									
										vendored
									
									
								
							
							
						
						
									
										58
									
								
								dist/merge/index.js
									
										
									
									
										vendored
									
									
								
							| 
						 | 
					@ -2997,7 +2997,7 @@ var __importDefault = (this && this.__importDefault) || function (mod) {
 | 
				
			||||||
    return (mod && mod.__esModule) ? mod : { "default": mod };
 | 
					    return (mod && mod.__esModule) ? mod : { "default": mod };
 | 
				
			||||||
};
 | 
					};
 | 
				
			||||||
Object.defineProperty(exports, "__esModule", ({ value: true }));
 | 
					Object.defineProperty(exports, "__esModule", ({ value: true }));
 | 
				
			||||||
exports.getConcurrency = exports.getGitHubWorkspaceDir = exports.isGhes = exports.getResultsServiceUrl = exports.getRuntimeToken = exports.getUploadChunkSize = void 0;
 | 
					exports.getUploadChunkTimeout = exports.getConcurrency = exports.getGitHubWorkspaceDir = exports.isGhes = exports.getResultsServiceUrl = exports.getRuntimeToken = exports.getUploadChunkSize = void 0;
 | 
				
			||||||
const os_1 = __importDefault(__nccwpck_require__(22037));
 | 
					const os_1 = __importDefault(__nccwpck_require__(22037));
 | 
				
			||||||
// Used for controlling the highWaterMark value of the zip that is being streamed
 | 
					// Used for controlling the highWaterMark value of the zip that is being streamed
 | 
				
			||||||
// The same value is used as the chunk size that is use during upload to blob storage
 | 
					// The same value is used as the chunk size that is use during upload to blob storage
 | 
				
			||||||
| 
						 | 
					@ -3050,6 +3050,10 @@ function getConcurrency() {
 | 
				
			||||||
    return concurrency > 300 ? 300 : concurrency;
 | 
					    return concurrency > 300 ? 300 : concurrency;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
exports.getConcurrency = getConcurrency;
 | 
					exports.getConcurrency = getConcurrency;
 | 
				
			||||||
 | 
					function getUploadChunkTimeout() {
 | 
				
			||||||
 | 
					    return 30000; // 30 seconds
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					exports.getUploadChunkTimeout = getUploadChunkTimeout;
 | 
				
			||||||
//# sourceMappingURL=config.js.map
 | 
					//# sourceMappingURL=config.js.map
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/***/ }),
 | 
					/***/ }),
 | 
				
			||||||
| 
						 | 
					@ -3298,37 +3302,34 @@ function uploadZipToBlobStorage(authenticatedUploadURL, zipUploadStream) {
 | 
				
			||||||
    return __awaiter(this, void 0, void 0, function* () {
 | 
					    return __awaiter(this, void 0, void 0, function* () {
 | 
				
			||||||
        let uploadByteCount = 0;
 | 
					        let uploadByteCount = 0;
 | 
				
			||||||
        let lastProgressTime = Date.now();
 | 
					        let lastProgressTime = Date.now();
 | 
				
			||||||
        let timeoutId;
 | 
					        const abortController = new AbortController();
 | 
				
			||||||
        const chunkTimer = (timeout) => {
 | 
					        const chunkTimer = (interval) => __awaiter(this, void 0, void 0, function* () {
 | 
				
			||||||
            // clear the previous timeout
 | 
					            return new Promise((resolve, reject) => {
 | 
				
			||||||
            if (timeoutId) {
 | 
					                const timer = setInterval(() => {
 | 
				
			||||||
                clearTimeout(timeoutId);
 | 
					                    if (Date.now() - lastProgressTime > interval) {
 | 
				
			||||||
            }
 | 
					                        reject(new Error('Upload progress stalled.'));
 | 
				
			||||||
            timeoutId = setTimeout(() => {
 | 
					                    }
 | 
				
			||||||
                const now = Date.now();
 | 
					                }, interval);
 | 
				
			||||||
                // if there's been more than 30 seconds since the
 | 
					                abortController.signal.addEventListener('abort', () => {
 | 
				
			||||||
                // last progress event, then we'll consider the upload stalled
 | 
					                    clearInterval(timer);
 | 
				
			||||||
                if (now - lastProgressTime > timeout) {
 | 
					                    resolve();
 | 
				
			||||||
                    throw new Error('Upload progress stalled.');
 | 
					                });
 | 
				
			||||||
                }
 | 
					            });
 | 
				
			||||||
            }, timeout);
 | 
					        });
 | 
				
			||||||
            return timeoutId;
 | 
					 | 
				
			||||||
        };
 | 
					 | 
				
			||||||
        const maxConcurrency = (0, config_1.getConcurrency)();
 | 
					        const maxConcurrency = (0, config_1.getConcurrency)();
 | 
				
			||||||
        const bufferSize = (0, config_1.getUploadChunkSize)();
 | 
					        const bufferSize = (0, config_1.getUploadChunkSize)();
 | 
				
			||||||
        const blobClient = new storage_blob_1.BlobClient(authenticatedUploadURL);
 | 
					        const blobClient = new storage_blob_1.BlobClient(authenticatedUploadURL);
 | 
				
			||||||
        const blockBlobClient = blobClient.getBlockBlobClient();
 | 
					        const blockBlobClient = blobClient.getBlockBlobClient();
 | 
				
			||||||
        const timeoutDuration = 300000; // 30 seconds
 | 
					 | 
				
			||||||
        core.debug(`Uploading artifact zip to blob storage with maxConcurrency: ${maxConcurrency}, bufferSize: ${bufferSize}`);
 | 
					        core.debug(`Uploading artifact zip to blob storage with maxConcurrency: ${maxConcurrency}, bufferSize: ${bufferSize}`);
 | 
				
			||||||
        const uploadCallback = (progress) => {
 | 
					        const uploadCallback = (progress) => {
 | 
				
			||||||
            core.info(`Uploaded bytes ${progress.loadedBytes}`);
 | 
					            core.info(`Uploaded bytes ${progress.loadedBytes}`);
 | 
				
			||||||
            uploadByteCount = progress.loadedBytes;
 | 
					            uploadByteCount = progress.loadedBytes;
 | 
				
			||||||
            chunkTimer(timeoutDuration);
 | 
					 | 
				
			||||||
            lastProgressTime = Date.now();
 | 
					            lastProgressTime = Date.now();
 | 
				
			||||||
        };
 | 
					        };
 | 
				
			||||||
        const options = {
 | 
					        const options = {
 | 
				
			||||||
            blobHTTPHeaders: { blobContentType: 'zip' },
 | 
					            blobHTTPHeaders: { blobContentType: 'zip' },
 | 
				
			||||||
            onProgress: uploadCallback
 | 
					            onProgress: uploadCallback,
 | 
				
			||||||
 | 
					            abortSignal: abortController.signal
 | 
				
			||||||
        };
 | 
					        };
 | 
				
			||||||
        let sha256Hash = undefined;
 | 
					        let sha256Hash = undefined;
 | 
				
			||||||
        const uploadStream = new stream.PassThrough();
 | 
					        const uploadStream = new stream.PassThrough();
 | 
				
			||||||
| 
						 | 
					@ -3337,9 +3338,10 @@ function uploadZipToBlobStorage(authenticatedUploadURL, zipUploadStream) {
 | 
				
			||||||
        zipUploadStream.pipe(hashStream).setEncoding('hex'); // This stream is used to compute a hash of the zip content that gets used. Integrity check
 | 
					        zipUploadStream.pipe(hashStream).setEncoding('hex'); // This stream is used to compute a hash of the zip content that gets used. Integrity check
 | 
				
			||||||
        core.info('Beginning upload of artifact content to blob storage');
 | 
					        core.info('Beginning upload of artifact content to blob storage');
 | 
				
			||||||
        try {
 | 
					        try {
 | 
				
			||||||
            // Start the chunk timer
 | 
					            yield Promise.race([
 | 
				
			||||||
            timeoutId = chunkTimer(timeoutDuration);
 | 
					                blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options),
 | 
				
			||||||
            yield blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options);
 | 
					                chunkTimer((0, config_1.getUploadChunkTimeout)())
 | 
				
			||||||
 | 
					            ]);
 | 
				
			||||||
        }
 | 
					        }
 | 
				
			||||||
        catch (error) {
 | 
					        catch (error) {
 | 
				
			||||||
            if (errors_1.NetworkError.isNetworkErrorCode(error === null || error === void 0 ? void 0 : error.code)) {
 | 
					            if (errors_1.NetworkError.isNetworkErrorCode(error === null || error === void 0 ? void 0 : error.code)) {
 | 
				
			||||||
| 
						 | 
					@ -3348,10 +3350,7 @@ function uploadZipToBlobStorage(authenticatedUploadURL, zipUploadStream) {
 | 
				
			||||||
            throw error;
 | 
					            throw error;
 | 
				
			||||||
        }
 | 
					        }
 | 
				
			||||||
        finally {
 | 
					        finally {
 | 
				
			||||||
            // clear the timeout whether or not the upload completes
 | 
					            abortController.abort();
 | 
				
			||||||
            if (timeoutId) {
 | 
					 | 
				
			||||||
                clearTimeout(timeoutId);
 | 
					 | 
				
			||||||
            }
 | 
					 | 
				
			||||||
        }
 | 
					        }
 | 
				
			||||||
        core.info('Finished uploading artifact content to blob storage!');
 | 
					        core.info('Finished uploading artifact content to blob storage!');
 | 
				
			||||||
        hashStream.end();
 | 
					        hashStream.end();
 | 
				
			||||||
| 
						 | 
					@ -3778,7 +3777,6 @@ exports.createZipUploadStream = exports.ZipUploadStream = exports.DEFAULT_COMPRE
 | 
				
			||||||
const stream = __importStar(__nccwpck_require__(12781));
 | 
					const stream = __importStar(__nccwpck_require__(12781));
 | 
				
			||||||
const archiver = __importStar(__nccwpck_require__(43084));
 | 
					const archiver = __importStar(__nccwpck_require__(43084));
 | 
				
			||||||
const core = __importStar(__nccwpck_require__(42186));
 | 
					const core = __importStar(__nccwpck_require__(42186));
 | 
				
			||||||
const fs_1 = __nccwpck_require__(57147);
 | 
					 | 
				
			||||||
const config_1 = __nccwpck_require__(74610);
 | 
					const config_1 = __nccwpck_require__(74610);
 | 
				
			||||||
exports.DEFAULT_COMPRESSION_LEVEL = 6;
 | 
					exports.DEFAULT_COMPRESSION_LEVEL = 6;
 | 
				
			||||||
// Custom stream transformer so we can set the highWaterMark property
 | 
					// Custom stream transformer so we can set the highWaterMark property
 | 
				
			||||||
| 
						 | 
					@ -3810,7 +3808,7 @@ function createZipUploadStream(uploadSpecification, compressionLevel = exports.D
 | 
				
			||||||
        for (const file of uploadSpecification) {
 | 
					        for (const file of uploadSpecification) {
 | 
				
			||||||
            if (file.sourcePath !== null) {
 | 
					            if (file.sourcePath !== null) {
 | 
				
			||||||
                // Add a normal file to the zip
 | 
					                // Add a normal file to the zip
 | 
				
			||||||
                zip.append((0, fs_1.createReadStream)(file.sourcePath), {
 | 
					                zip.file(file.sourcePath, {
 | 
				
			||||||
                    name: file.destinationPath
 | 
					                    name: file.destinationPath
 | 
				
			||||||
                });
 | 
					                });
 | 
				
			||||||
            }
 | 
					            }
 | 
				
			||||||
| 
						 | 
					@ -136152,7 +136150,7 @@ module.exports = index;
 | 
				
			||||||
/***/ ((module) => {
 | 
					/***/ ((module) => {
 | 
				
			||||||
 | 
					
 | 
				
			||||||
"use strict";
 | 
					"use strict";
 | 
				
			||||||
module.exports = JSON.parse('{"name":"@actions/artifact","version":"2.1.8","preview":true,"description":"Actions artifact lib","keywords":["github","actions","artifact"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/artifact","license":"MIT","main":"lib/artifact.js","types":"lib/artifact.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/artifact"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"cd ../../ && npm run test ./packages/artifact","bootstrap":"cd ../../ && npm run bootstrap","tsc-run":"tsc","tsc":"npm run bootstrap && npm run tsc-run","gen:docs":"typedoc --plugin typedoc-plugin-markdown --out docs/generated src/artifact.ts --githubPages false --readme none"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.10.0","@actions/github":"^5.1.1","@actions/http-client":"^2.1.0","@azure/storage-blob":"^12.15.0","@octokit/core":"^3.5.1","@octokit/plugin-request-log":"^1.0.4","@octokit/plugin-retry":"^3.0.9","@octokit/request-error":"^5.0.0","@protobuf-ts/plugin":"^2.2.3-alpha.1","archiver":"^7.0.1","crypto":"^1.0.1","jwt-decode":"^3.1.2","twirp-ts":"^2.5.0","unzip-stream":"^0.3.1"},"devDependencies":{"@types/archiver":"^5.3.2","@types/unzip-stream":"^0.3.4","typedoc":"^0.25.4","typedoc-plugin-markdown":"^3.17.1","typescript":"^5.2.2"}}');
 | 
					module.exports = JSON.parse('{"name":"@actions/artifact","version":"2.1.9","preview":true,"description":"Actions artifact lib","keywords":["github","actions","artifact"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/artifact","license":"MIT","main":"lib/artifact.js","types":"lib/artifact.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/artifact"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"cd ../../ && npm run test ./packages/artifact","bootstrap":"cd ../../ && npm run bootstrap","tsc-run":"tsc","tsc":"npm run bootstrap && npm run tsc-run","gen:docs":"typedoc --plugin typedoc-plugin-markdown --out docs/generated src/artifact.ts --githubPages false --readme none"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.10.0","@actions/github":"^5.1.1","@actions/http-client":"^2.1.0","@azure/storage-blob":"^12.15.0","@octokit/core":"^3.5.1","@octokit/plugin-request-log":"^1.0.4","@octokit/plugin-retry":"^3.0.9","@octokit/request-error":"^5.0.0","@protobuf-ts/plugin":"^2.2.3-alpha.1","archiver":"^7.0.1","crypto":"^1.0.1","jwt-decode":"^3.1.2","twirp-ts":"^2.5.0","unzip-stream":"^0.3.1"},"devDependencies":{"@types/archiver":"^5.3.2","@types/unzip-stream":"^0.3.4","typedoc":"^0.25.4","typedoc-plugin-markdown":"^3.17.1","typescript":"^5.2.2"}}');
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/***/ }),
 | 
					/***/ }),
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
							
								
								
									
										58
									
								
								dist/upload/index.js
									
										
									
									
										vendored
									
									
								
							
							
						
						
									
										58
									
								
								dist/upload/index.js
									
										
									
									
										vendored
									
									
								
							| 
						 | 
					@ -2997,7 +2997,7 @@ var __importDefault = (this && this.__importDefault) || function (mod) {
 | 
				
			||||||
    return (mod && mod.__esModule) ? mod : { "default": mod };
 | 
					    return (mod && mod.__esModule) ? mod : { "default": mod };
 | 
				
			||||||
};
 | 
					};
 | 
				
			||||||
Object.defineProperty(exports, "__esModule", ({ value: true }));
 | 
					Object.defineProperty(exports, "__esModule", ({ value: true }));
 | 
				
			||||||
exports.getConcurrency = exports.getGitHubWorkspaceDir = exports.isGhes = exports.getResultsServiceUrl = exports.getRuntimeToken = exports.getUploadChunkSize = void 0;
 | 
					exports.getUploadChunkTimeout = exports.getConcurrency = exports.getGitHubWorkspaceDir = exports.isGhes = exports.getResultsServiceUrl = exports.getRuntimeToken = exports.getUploadChunkSize = void 0;
 | 
				
			||||||
const os_1 = __importDefault(__nccwpck_require__(22037));
 | 
					const os_1 = __importDefault(__nccwpck_require__(22037));
 | 
				
			||||||
// Used for controlling the highWaterMark value of the zip that is being streamed
 | 
					// Used for controlling the highWaterMark value of the zip that is being streamed
 | 
				
			||||||
// The same value is used as the chunk size that is use during upload to blob storage
 | 
					// The same value is used as the chunk size that is use during upload to blob storage
 | 
				
			||||||
| 
						 | 
					@ -3050,6 +3050,10 @@ function getConcurrency() {
 | 
				
			||||||
    return concurrency > 300 ? 300 : concurrency;
 | 
					    return concurrency > 300 ? 300 : concurrency;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
exports.getConcurrency = getConcurrency;
 | 
					exports.getConcurrency = getConcurrency;
 | 
				
			||||||
 | 
					function getUploadChunkTimeout() {
 | 
				
			||||||
 | 
					    return 30000; // 30 seconds
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					exports.getUploadChunkTimeout = getUploadChunkTimeout;
 | 
				
			||||||
//# sourceMappingURL=config.js.map
 | 
					//# sourceMappingURL=config.js.map
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/***/ }),
 | 
					/***/ }),
 | 
				
			||||||
| 
						 | 
					@ -3298,37 +3302,34 @@ function uploadZipToBlobStorage(authenticatedUploadURL, zipUploadStream) {
 | 
				
			||||||
    return __awaiter(this, void 0, void 0, function* () {
 | 
					    return __awaiter(this, void 0, void 0, function* () {
 | 
				
			||||||
        let uploadByteCount = 0;
 | 
					        let uploadByteCount = 0;
 | 
				
			||||||
        let lastProgressTime = Date.now();
 | 
					        let lastProgressTime = Date.now();
 | 
				
			||||||
        let timeoutId;
 | 
					        const abortController = new AbortController();
 | 
				
			||||||
        const chunkTimer = (timeout) => {
 | 
					        const chunkTimer = (interval) => __awaiter(this, void 0, void 0, function* () {
 | 
				
			||||||
            // clear the previous timeout
 | 
					            return new Promise((resolve, reject) => {
 | 
				
			||||||
            if (timeoutId) {
 | 
					                const timer = setInterval(() => {
 | 
				
			||||||
                clearTimeout(timeoutId);
 | 
					                    if (Date.now() - lastProgressTime > interval) {
 | 
				
			||||||
            }
 | 
					                        reject(new Error('Upload progress stalled.'));
 | 
				
			||||||
            timeoutId = setTimeout(() => {
 | 
					                    }
 | 
				
			||||||
                const now = Date.now();
 | 
					                }, interval);
 | 
				
			||||||
                // if there's been more than 30 seconds since the
 | 
					                abortController.signal.addEventListener('abort', () => {
 | 
				
			||||||
                // last progress event, then we'll consider the upload stalled
 | 
					                    clearInterval(timer);
 | 
				
			||||||
                if (now - lastProgressTime > timeout) {
 | 
					                    resolve();
 | 
				
			||||||
                    throw new Error('Upload progress stalled.');
 | 
					                });
 | 
				
			||||||
                }
 | 
					            });
 | 
				
			||||||
            }, timeout);
 | 
					        });
 | 
				
			||||||
            return timeoutId;
 | 
					 | 
				
			||||||
        };
 | 
					 | 
				
			||||||
        const maxConcurrency = (0, config_1.getConcurrency)();
 | 
					        const maxConcurrency = (0, config_1.getConcurrency)();
 | 
				
			||||||
        const bufferSize = (0, config_1.getUploadChunkSize)();
 | 
					        const bufferSize = (0, config_1.getUploadChunkSize)();
 | 
				
			||||||
        const blobClient = new storage_blob_1.BlobClient(authenticatedUploadURL);
 | 
					        const blobClient = new storage_blob_1.BlobClient(authenticatedUploadURL);
 | 
				
			||||||
        const blockBlobClient = blobClient.getBlockBlobClient();
 | 
					        const blockBlobClient = blobClient.getBlockBlobClient();
 | 
				
			||||||
        const timeoutDuration = 300000; // 30 seconds
 | 
					 | 
				
			||||||
        core.debug(`Uploading artifact zip to blob storage with maxConcurrency: ${maxConcurrency}, bufferSize: ${bufferSize}`);
 | 
					        core.debug(`Uploading artifact zip to blob storage with maxConcurrency: ${maxConcurrency}, bufferSize: ${bufferSize}`);
 | 
				
			||||||
        const uploadCallback = (progress) => {
 | 
					        const uploadCallback = (progress) => {
 | 
				
			||||||
            core.info(`Uploaded bytes ${progress.loadedBytes}`);
 | 
					            core.info(`Uploaded bytes ${progress.loadedBytes}`);
 | 
				
			||||||
            uploadByteCount = progress.loadedBytes;
 | 
					            uploadByteCount = progress.loadedBytes;
 | 
				
			||||||
            chunkTimer(timeoutDuration);
 | 
					 | 
				
			||||||
            lastProgressTime = Date.now();
 | 
					            lastProgressTime = Date.now();
 | 
				
			||||||
        };
 | 
					        };
 | 
				
			||||||
        const options = {
 | 
					        const options = {
 | 
				
			||||||
            blobHTTPHeaders: { blobContentType: 'zip' },
 | 
					            blobHTTPHeaders: { blobContentType: 'zip' },
 | 
				
			||||||
            onProgress: uploadCallback
 | 
					            onProgress: uploadCallback,
 | 
				
			||||||
 | 
					            abortSignal: abortController.signal
 | 
				
			||||||
        };
 | 
					        };
 | 
				
			||||||
        let sha256Hash = undefined;
 | 
					        let sha256Hash = undefined;
 | 
				
			||||||
        const uploadStream = new stream.PassThrough();
 | 
					        const uploadStream = new stream.PassThrough();
 | 
				
			||||||
| 
						 | 
					@ -3337,9 +3338,10 @@ function uploadZipToBlobStorage(authenticatedUploadURL, zipUploadStream) {
 | 
				
			||||||
        zipUploadStream.pipe(hashStream).setEncoding('hex'); // This stream is used to compute a hash of the zip content that gets used. Integrity check
 | 
					        zipUploadStream.pipe(hashStream).setEncoding('hex'); // This stream is used to compute a hash of the zip content that gets used. Integrity check
 | 
				
			||||||
        core.info('Beginning upload of artifact content to blob storage');
 | 
					        core.info('Beginning upload of artifact content to blob storage');
 | 
				
			||||||
        try {
 | 
					        try {
 | 
				
			||||||
            // Start the chunk timer
 | 
					            yield Promise.race([
 | 
				
			||||||
            timeoutId = chunkTimer(timeoutDuration);
 | 
					                blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options),
 | 
				
			||||||
            yield blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options);
 | 
					                chunkTimer((0, config_1.getUploadChunkTimeout)())
 | 
				
			||||||
 | 
					            ]);
 | 
				
			||||||
        }
 | 
					        }
 | 
				
			||||||
        catch (error) {
 | 
					        catch (error) {
 | 
				
			||||||
            if (errors_1.NetworkError.isNetworkErrorCode(error === null || error === void 0 ? void 0 : error.code)) {
 | 
					            if (errors_1.NetworkError.isNetworkErrorCode(error === null || error === void 0 ? void 0 : error.code)) {
 | 
				
			||||||
| 
						 | 
					@ -3348,10 +3350,7 @@ function uploadZipToBlobStorage(authenticatedUploadURL, zipUploadStream) {
 | 
				
			||||||
            throw error;
 | 
					            throw error;
 | 
				
			||||||
        }
 | 
					        }
 | 
				
			||||||
        finally {
 | 
					        finally {
 | 
				
			||||||
            // clear the timeout whether or not the upload completes
 | 
					            abortController.abort();
 | 
				
			||||||
            if (timeoutId) {
 | 
					 | 
				
			||||||
                clearTimeout(timeoutId);
 | 
					 | 
				
			||||||
            }
 | 
					 | 
				
			||||||
        }
 | 
					        }
 | 
				
			||||||
        core.info('Finished uploading artifact content to blob storage!');
 | 
					        core.info('Finished uploading artifact content to blob storage!');
 | 
				
			||||||
        hashStream.end();
 | 
					        hashStream.end();
 | 
				
			||||||
| 
						 | 
					@ -3778,7 +3777,6 @@ exports.createZipUploadStream = exports.ZipUploadStream = exports.DEFAULT_COMPRE
 | 
				
			||||||
const stream = __importStar(__nccwpck_require__(12781));
 | 
					const stream = __importStar(__nccwpck_require__(12781));
 | 
				
			||||||
const archiver = __importStar(__nccwpck_require__(43084));
 | 
					const archiver = __importStar(__nccwpck_require__(43084));
 | 
				
			||||||
const core = __importStar(__nccwpck_require__(42186));
 | 
					const core = __importStar(__nccwpck_require__(42186));
 | 
				
			||||||
const fs_1 = __nccwpck_require__(57147);
 | 
					 | 
				
			||||||
const config_1 = __nccwpck_require__(74610);
 | 
					const config_1 = __nccwpck_require__(74610);
 | 
				
			||||||
exports.DEFAULT_COMPRESSION_LEVEL = 6;
 | 
					exports.DEFAULT_COMPRESSION_LEVEL = 6;
 | 
				
			||||||
// Custom stream transformer so we can set the highWaterMark property
 | 
					// Custom stream transformer so we can set the highWaterMark property
 | 
				
			||||||
| 
						 | 
					@ -3810,7 +3808,7 @@ function createZipUploadStream(uploadSpecification, compressionLevel = exports.D
 | 
				
			||||||
        for (const file of uploadSpecification) {
 | 
					        for (const file of uploadSpecification) {
 | 
				
			||||||
            if (file.sourcePath !== null) {
 | 
					            if (file.sourcePath !== null) {
 | 
				
			||||||
                // Add a normal file to the zip
 | 
					                // Add a normal file to the zip
 | 
				
			||||||
                zip.append((0, fs_1.createReadStream)(file.sourcePath), {
 | 
					                zip.file(file.sourcePath, {
 | 
				
			||||||
                    name: file.destinationPath
 | 
					                    name: file.destinationPath
 | 
				
			||||||
                });
 | 
					                });
 | 
				
			||||||
            }
 | 
					            }
 | 
				
			||||||
| 
						 | 
					@ -136162,7 +136160,7 @@ module.exports = index;
 | 
				
			||||||
/***/ ((module) => {
 | 
					/***/ ((module) => {
 | 
				
			||||||
 | 
					
 | 
				
			||||||
"use strict";
 | 
					"use strict";
 | 
				
			||||||
module.exports = JSON.parse('{"name":"@actions/artifact","version":"2.1.8","preview":true,"description":"Actions artifact lib","keywords":["github","actions","artifact"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/artifact","license":"MIT","main":"lib/artifact.js","types":"lib/artifact.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/artifact"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"cd ../../ && npm run test ./packages/artifact","bootstrap":"cd ../../ && npm run bootstrap","tsc-run":"tsc","tsc":"npm run bootstrap && npm run tsc-run","gen:docs":"typedoc --plugin typedoc-plugin-markdown --out docs/generated src/artifact.ts --githubPages false --readme none"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.10.0","@actions/github":"^5.1.1","@actions/http-client":"^2.1.0","@azure/storage-blob":"^12.15.0","@octokit/core":"^3.5.1","@octokit/plugin-request-log":"^1.0.4","@octokit/plugin-retry":"^3.0.9","@octokit/request-error":"^5.0.0","@protobuf-ts/plugin":"^2.2.3-alpha.1","archiver":"^7.0.1","crypto":"^1.0.1","jwt-decode":"^3.1.2","twirp-ts":"^2.5.0","unzip-stream":"^0.3.1"},"devDependencies":{"@types/archiver":"^5.3.2","@types/unzip-stream":"^0.3.4","typedoc":"^0.25.4","typedoc-plugin-markdown":"^3.17.1","typescript":"^5.2.2"}}');
 | 
					module.exports = JSON.parse('{"name":"@actions/artifact","version":"2.1.9","preview":true,"description":"Actions artifact lib","keywords":["github","actions","artifact"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/artifact","license":"MIT","main":"lib/artifact.js","types":"lib/artifact.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/artifact"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"cd ../../ && npm run test ./packages/artifact","bootstrap":"cd ../../ && npm run bootstrap","tsc-run":"tsc","tsc":"npm run bootstrap && npm run tsc-run","gen:docs":"typedoc --plugin typedoc-plugin-markdown --out docs/generated src/artifact.ts --githubPages false --readme none"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.10.0","@actions/github":"^5.1.1","@actions/http-client":"^2.1.0","@azure/storage-blob":"^12.15.0","@octokit/core":"^3.5.1","@octokit/plugin-request-log":"^1.0.4","@octokit/plugin-retry":"^3.0.9","@octokit/request-error":"^5.0.0","@protobuf-ts/plugin":"^2.2.3-alpha.1","archiver":"^7.0.1","crypto":"^1.0.1","jwt-decode":"^3.1.2","twirp-ts":"^2.5.0","unzip-stream":"^0.3.1"},"devDependencies":{"@types/archiver":"^5.3.2","@types/unzip-stream":"^0.3.4","typedoc":"^0.25.4","typedoc-plugin-markdown":"^3.17.1","typescript":"^5.2.2"}}');
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/***/ }),
 | 
					/***/ }),
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
							
								
								
									
										18
									
								
								package-lock.json
									
										
									
										generated
									
									
									
								
							
							
						
						
									
										18
									
								
								package-lock.json
									
										
									
										generated
									
									
									
								
							| 
						 | 
					@ -1,15 +1,15 @@
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
  "name": "upload-artifact",
 | 
					  "name": "upload-artifact",
 | 
				
			||||||
  "version": "4.3.4",
 | 
					  "version": "4.3.5",
 | 
				
			||||||
  "lockfileVersion": 2,
 | 
					  "lockfileVersion": 2,
 | 
				
			||||||
  "requires": true,
 | 
					  "requires": true,
 | 
				
			||||||
  "packages": {
 | 
					  "packages": {
 | 
				
			||||||
    "": {
 | 
					    "": {
 | 
				
			||||||
      "name": "upload-artifact",
 | 
					      "name": "upload-artifact",
 | 
				
			||||||
      "version": "4.3.4",
 | 
					      "version": "4.3.5",
 | 
				
			||||||
      "license": "MIT",
 | 
					      "license": "MIT",
 | 
				
			||||||
      "dependencies": {
 | 
					      "dependencies": {
 | 
				
			||||||
        "@actions/artifact": "^2.1.8",
 | 
					        "@actions/artifact": "^2.1.9",
 | 
				
			||||||
        "@actions/core": "^1.10.1",
 | 
					        "@actions/core": "^1.10.1",
 | 
				
			||||||
        "@actions/github": "^6.0.0",
 | 
					        "@actions/github": "^6.0.0",
 | 
				
			||||||
        "@actions/glob": "^0.3.0",
 | 
					        "@actions/glob": "^0.3.0",
 | 
				
			||||||
| 
						 | 
					@ -34,9 +34,9 @@
 | 
				
			||||||
      }
 | 
					      }
 | 
				
			||||||
    },
 | 
					    },
 | 
				
			||||||
    "node_modules/@actions/artifact": {
 | 
					    "node_modules/@actions/artifact": {
 | 
				
			||||||
      "version": "2.1.8",
 | 
					      "version": "2.1.9",
 | 
				
			||||||
      "resolved": "https://registry.npmjs.org/@actions/artifact/-/artifact-2.1.8.tgz",
 | 
					      "resolved": "https://registry.npmjs.org/@actions/artifact/-/artifact-2.1.9.tgz",
 | 
				
			||||||
      "integrity": "sha512-kxgbllgF5f6mEdMeSW6WXlUbV1U77V9ECpA7LOYaY+Tm6RfXOm36EdXbpm+T9VPeaVqXK4QHLAgqay9GSyClgw==",
 | 
					      "integrity": "sha512-f9JXC9JrwramDRJHZiIOKJo3PGw/V3riYegLj5kHi8YEJ2k72TNUd1zDW1BG50ILnzJ0cp1faDVJ2pSdolRQfg==",
 | 
				
			||||||
      "dependencies": {
 | 
					      "dependencies": {
 | 
				
			||||||
        "@actions/core": "^1.10.0",
 | 
					        "@actions/core": "^1.10.0",
 | 
				
			||||||
        "@actions/github": "^5.1.1",
 | 
					        "@actions/github": "^5.1.1",
 | 
				
			||||||
| 
						 | 
					@ -7902,9 +7902,9 @@
 | 
				
			||||||
  },
 | 
					  },
 | 
				
			||||||
  "dependencies": {
 | 
					  "dependencies": {
 | 
				
			||||||
    "@actions/artifact": {
 | 
					    "@actions/artifact": {
 | 
				
			||||||
      "version": "2.1.8",
 | 
					      "version": "2.1.9",
 | 
				
			||||||
      "resolved": "https://registry.npmjs.org/@actions/artifact/-/artifact-2.1.8.tgz",
 | 
					      "resolved": "https://registry.npmjs.org/@actions/artifact/-/artifact-2.1.9.tgz",
 | 
				
			||||||
      "integrity": "sha512-kxgbllgF5f6mEdMeSW6WXlUbV1U77V9ECpA7LOYaY+Tm6RfXOm36EdXbpm+T9VPeaVqXK4QHLAgqay9GSyClgw==",
 | 
					      "integrity": "sha512-f9JXC9JrwramDRJHZiIOKJo3PGw/V3riYegLj5kHi8YEJ2k72TNUd1zDW1BG50ILnzJ0cp1faDVJ2pSdolRQfg==",
 | 
				
			||||||
      "requires": {
 | 
					      "requires": {
 | 
				
			||||||
        "@actions/core": "^1.10.0",
 | 
					        "@actions/core": "^1.10.0",
 | 
				
			||||||
        "@actions/github": "^5.1.1",
 | 
					        "@actions/github": "^5.1.1",
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1,6 +1,6 @@
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
  "name": "upload-artifact",
 | 
					  "name": "upload-artifact",
 | 
				
			||||||
  "version": "4.3.4",
 | 
					  "version": "4.3.5",
 | 
				
			||||||
  "description": "Upload an Actions Artifact in a workflow run",
 | 
					  "description": "Upload an Actions Artifact in a workflow run",
 | 
				
			||||||
  "main": "dist/upload/index.js",
 | 
					  "main": "dist/upload/index.js",
 | 
				
			||||||
  "scripts": {
 | 
					  "scripts": {
 | 
				
			||||||
| 
						 | 
					@ -29,7 +29,7 @@
 | 
				
			||||||
  },
 | 
					  },
 | 
				
			||||||
  "homepage": "https://github.com/actions/upload-artifact#readme",
 | 
					  "homepage": "https://github.com/actions/upload-artifact#readme",
 | 
				
			||||||
  "dependencies": {
 | 
					  "dependencies": {
 | 
				
			||||||
    "@actions/artifact": "^2.1.8",
 | 
					    "@actions/artifact": "^2.1.9",
 | 
				
			||||||
    "@actions/core": "^1.10.1",
 | 
					    "@actions/core": "^1.10.1",
 | 
				
			||||||
    "@actions/github": "^6.0.0",
 | 
					    "@actions/github": "^6.0.0",
 | 
				
			||||||
    "@actions/glob": "^0.3.0",
 | 
					    "@actions/glob": "^0.3.0",
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
	Add table
		Add a link
		
	
		Reference in a new issue