mirror of
https://github.com/actions/upload-artifact.git
synced 2025-04-22 19:11:21 +00:00
13071 lines
395 KiB
JavaScript
13071 lines
395 KiB
JavaScript
module.exports =
|
||
/******/ (function(modules, runtime) { // webpackBootstrap
|
||
/******/ "use strict";
|
||
/******/ // The module cache
|
||
/******/ var installedModules = {};
|
||
/******/
|
||
/******/ // The require function
|
||
/******/ function __webpack_require__(moduleId) {
|
||
/******/
|
||
/******/ // Check if module is in cache
|
||
/******/ if(installedModules[moduleId]) {
|
||
/******/ return installedModules[moduleId].exports;
|
||
/******/ }
|
||
/******/ // Create a new module (and put it into the cache)
|
||
/******/ var module = installedModules[moduleId] = {
|
||
/******/ i: moduleId,
|
||
/******/ l: false,
|
||
/******/ exports: {}
|
||
/******/ };
|
||
/******/
|
||
/******/ // Execute the module function
|
||
/******/ modules[moduleId].call(module.exports, module, module.exports, __webpack_require__);
|
||
/******/
|
||
/******/ // Flag the module as loaded
|
||
/******/ module.l = true;
|
||
/******/
|
||
/******/ // Return the exports of the module
|
||
/******/ return module.exports;
|
||
/******/ }
|
||
/******/
|
||
/******/
|
||
/******/ __webpack_require__.ab = __dirname + "/";
|
||
/******/
|
||
/******/ // the startup function
|
||
/******/ function startup() {
|
||
/******/ // Load entry module and return exports
|
||
/******/ return __webpack_require__(385);
|
||
/******/ };
|
||
/******/
|
||
/******/ // run startup
|
||
/******/ return startup();
|
||
/******/ })
|
||
/************************************************************************/
|
||
/******/ ({
|
||
|
||
/***/ 11:
|
||
/***/ (function(module) {
|
||
|
||
// Returns a wrapper function that returns a wrapped callback
|
||
// The wrapper function should do some stuff, and return a
|
||
// presumably different callback function.
|
||
// This makes sure that own properties are retained, so that
|
||
// decorations and such are not lost along the way.
|
||
module.exports = wrappy
|
||
function wrappy (fn, cb) {
|
||
if (fn && cb) return wrappy(fn)(cb)
|
||
|
||
if (typeof fn !== 'function')
|
||
throw new TypeError('need wrapper function')
|
||
|
||
Object.keys(fn).forEach(function (k) {
|
||
wrapper[k] = fn[k]
|
||
})
|
||
|
||
return wrapper
|
||
|
||
function wrapper() {
|
||
var args = new Array(arguments.length)
|
||
for (var i = 0; i < args.length; i++) {
|
||
args[i] = arguments[i]
|
||
}
|
||
var ret = fn.apply(this, args)
|
||
var cb = args[args.length-1]
|
||
if (typeof ret === 'function' && ret !== cb) {
|
||
Object.keys(cb).forEach(function (k) {
|
||
ret[k] = cb[k]
|
||
})
|
||
}
|
||
return ret
|
||
}
|
||
}
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 16:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("tls");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 26:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var enhanceError = __webpack_require__(369);
|
||
|
||
/**
|
||
* Create an Error with the specified message, config, error code, request and response.
|
||
*
|
||
* @param {string} message The error message.
|
||
* @param {Object} config The config.
|
||
* @param {string} [code] The error code (for example, 'ECONNABORTED').
|
||
* @param {Object} [request] The request.
|
||
* @param {Object} [response] The response.
|
||
* @returns {Error} The created error.
|
||
*/
|
||
module.exports = function createError(message, config, code, request, response) {
|
||
var error = new Error(message);
|
||
return enhanceError(error, config, code, request, response);
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 35:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var bind = __webpack_require__(727);
|
||
|
||
// utils is a library of generic helper functions non-specific to axios
|
||
|
||
var toString = Object.prototype.toString;
|
||
|
||
/**
|
||
* Determine if a value is an Array
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if value is an Array, otherwise false
|
||
*/
|
||
function isArray(val) {
|
||
return toString.call(val) === '[object Array]';
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is undefined
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if the value is undefined, otherwise false
|
||
*/
|
||
function isUndefined(val) {
|
||
return typeof val === 'undefined';
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is a Buffer
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if value is a Buffer, otherwise false
|
||
*/
|
||
function isBuffer(val) {
|
||
return val !== null && !isUndefined(val) && val.constructor !== null && !isUndefined(val.constructor)
|
||
&& typeof val.constructor.isBuffer === 'function' && val.constructor.isBuffer(val);
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is an ArrayBuffer
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if value is an ArrayBuffer, otherwise false
|
||
*/
|
||
function isArrayBuffer(val) {
|
||
return toString.call(val) === '[object ArrayBuffer]';
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is a FormData
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if value is an FormData, otherwise false
|
||
*/
|
||
function isFormData(val) {
|
||
return (typeof FormData !== 'undefined') && (val instanceof FormData);
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is a view on an ArrayBuffer
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if value is a view on an ArrayBuffer, otherwise false
|
||
*/
|
||
function isArrayBufferView(val) {
|
||
var result;
|
||
if ((typeof ArrayBuffer !== 'undefined') && (ArrayBuffer.isView)) {
|
||
result = ArrayBuffer.isView(val);
|
||
} else {
|
||
result = (val) && (val.buffer) && (val.buffer instanceof ArrayBuffer);
|
||
}
|
||
return result;
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is a String
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if value is a String, otherwise false
|
||
*/
|
||
function isString(val) {
|
||
return typeof val === 'string';
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is a Number
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if value is a Number, otherwise false
|
||
*/
|
||
function isNumber(val) {
|
||
return typeof val === 'number';
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is an Object
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if value is an Object, otherwise false
|
||
*/
|
||
function isObject(val) {
|
||
return val !== null && typeof val === 'object';
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is a plain Object
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @return {boolean} True if value is a plain Object, otherwise false
|
||
*/
|
||
function isPlainObject(val) {
|
||
if (toString.call(val) !== '[object Object]') {
|
||
return false;
|
||
}
|
||
|
||
var prototype = Object.getPrototypeOf(val);
|
||
return prototype === null || prototype === Object.prototype;
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is a Date
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if value is a Date, otherwise false
|
||
*/
|
||
function isDate(val) {
|
||
return toString.call(val) === '[object Date]';
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is a File
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if value is a File, otherwise false
|
||
*/
|
||
function isFile(val) {
|
||
return toString.call(val) === '[object File]';
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is a Blob
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if value is a Blob, otherwise false
|
||
*/
|
||
function isBlob(val) {
|
||
return toString.call(val) === '[object Blob]';
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is a Function
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if value is a Function, otherwise false
|
||
*/
|
||
function isFunction(val) {
|
||
return toString.call(val) === '[object Function]';
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is a Stream
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if value is a Stream, otherwise false
|
||
*/
|
||
function isStream(val) {
|
||
return isObject(val) && isFunction(val.pipe);
|
||
}
|
||
|
||
/**
|
||
* Determine if a value is a URLSearchParams object
|
||
*
|
||
* @param {Object} val The value to test
|
||
* @returns {boolean} True if value is a URLSearchParams object, otherwise false
|
||
*/
|
||
function isURLSearchParams(val) {
|
||
return typeof URLSearchParams !== 'undefined' && val instanceof URLSearchParams;
|
||
}
|
||
|
||
/**
|
||
* Trim excess whitespace off the beginning and end of a string
|
||
*
|
||
* @param {String} str The String to trim
|
||
* @returns {String} The String freed of excess whitespace
|
||
*/
|
||
function trim(str) {
|
||
return str.trim ? str.trim() : str.replace(/^\s+|\s+$/g, '');
|
||
}
|
||
|
||
/**
|
||
* Determine if we're running in a standard browser environment
|
||
*
|
||
* This allows axios to run in a web worker, and react-native.
|
||
* Both environments support XMLHttpRequest, but not fully standard globals.
|
||
*
|
||
* web workers:
|
||
* typeof window -> undefined
|
||
* typeof document -> undefined
|
||
*
|
||
* react-native:
|
||
* navigator.product -> 'ReactNative'
|
||
* nativescript
|
||
* navigator.product -> 'NativeScript' or 'NS'
|
||
*/
|
||
function isStandardBrowserEnv() {
|
||
if (typeof navigator !== 'undefined' && (navigator.product === 'ReactNative' ||
|
||
navigator.product === 'NativeScript' ||
|
||
navigator.product === 'NS')) {
|
||
return false;
|
||
}
|
||
return (
|
||
typeof window !== 'undefined' &&
|
||
typeof document !== 'undefined'
|
||
);
|
||
}
|
||
|
||
/**
|
||
* Iterate over an Array or an Object invoking a function for each item.
|
||
*
|
||
* If `obj` is an Array callback will be called passing
|
||
* the value, index, and complete array for each item.
|
||
*
|
||
* If 'obj' is an Object callback will be called passing
|
||
* the value, key, and complete object for each property.
|
||
*
|
||
* @param {Object|Array} obj The object to iterate
|
||
* @param {Function} fn The callback to invoke for each item
|
||
*/
|
||
function forEach(obj, fn) {
|
||
// Don't bother if no value provided
|
||
if (obj === null || typeof obj === 'undefined') {
|
||
return;
|
||
}
|
||
|
||
// Force an array if not already something iterable
|
||
if (typeof obj !== 'object') {
|
||
/*eslint no-param-reassign:0*/
|
||
obj = [obj];
|
||
}
|
||
|
||
if (isArray(obj)) {
|
||
// Iterate over array values
|
||
for (var i = 0, l = obj.length; i < l; i++) {
|
||
fn.call(null, obj[i], i, obj);
|
||
}
|
||
} else {
|
||
// Iterate over object keys
|
||
for (var key in obj) {
|
||
if (Object.prototype.hasOwnProperty.call(obj, key)) {
|
||
fn.call(null, obj[key], key, obj);
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
/**
|
||
* Accepts varargs expecting each argument to be an object, then
|
||
* immutably merges the properties of each object and returns result.
|
||
*
|
||
* When multiple objects contain the same key the later object in
|
||
* the arguments list will take precedence.
|
||
*
|
||
* Example:
|
||
*
|
||
* ```js
|
||
* var result = merge({foo: 123}, {foo: 456});
|
||
* console.log(result.foo); // outputs 456
|
||
* ```
|
||
*
|
||
* @param {Object} obj1 Object to merge
|
||
* @returns {Object} Result of all merge properties
|
||
*/
|
||
function merge(/* obj1, obj2, obj3, ... */) {
|
||
var result = {};
|
||
function assignValue(val, key) {
|
||
if (isPlainObject(result[key]) && isPlainObject(val)) {
|
||
result[key] = merge(result[key], val);
|
||
} else if (isPlainObject(val)) {
|
||
result[key] = merge({}, val);
|
||
} else if (isArray(val)) {
|
||
result[key] = val.slice();
|
||
} else {
|
||
result[key] = val;
|
||
}
|
||
}
|
||
|
||
for (var i = 0, l = arguments.length; i < l; i++) {
|
||
forEach(arguments[i], assignValue);
|
||
}
|
||
return result;
|
||
}
|
||
|
||
/**
|
||
* Extends object a by mutably adding to it the properties of object b.
|
||
*
|
||
* @param {Object} a The object to be extended
|
||
* @param {Object} b The object to copy properties from
|
||
* @param {Object} thisArg The object to bind function to
|
||
* @return {Object} The resulting value of object a
|
||
*/
|
||
function extend(a, b, thisArg) {
|
||
forEach(b, function assignValue(val, key) {
|
||
if (thisArg && typeof val === 'function') {
|
||
a[key] = bind(val, thisArg);
|
||
} else {
|
||
a[key] = val;
|
||
}
|
||
});
|
||
return a;
|
||
}
|
||
|
||
/**
|
||
* Remove byte order marker. This catches EF BB BF (the UTF-8 BOM)
|
||
*
|
||
* @param {string} content with BOM
|
||
* @return {string} content value without BOM
|
||
*/
|
||
function stripBOM(content) {
|
||
if (content.charCodeAt(0) === 0xFEFF) {
|
||
content = content.slice(1);
|
||
}
|
||
return content;
|
||
}
|
||
|
||
module.exports = {
|
||
isArray: isArray,
|
||
isArrayBuffer: isArrayBuffer,
|
||
isBuffer: isBuffer,
|
||
isFormData: isFormData,
|
||
isArrayBufferView: isArrayBufferView,
|
||
isString: isString,
|
||
isNumber: isNumber,
|
||
isObject: isObject,
|
||
isPlainObject: isPlainObject,
|
||
isUndefined: isUndefined,
|
||
isDate: isDate,
|
||
isFile: isFile,
|
||
isBlob: isBlob,
|
||
isFunction: isFunction,
|
||
isStream: isStream,
|
||
isURLSearchParams: isURLSearchParams,
|
||
isStandardBrowserEnv: isStandardBrowserEnv,
|
||
forEach: forEach,
|
||
merge: merge,
|
||
extend: extend,
|
||
trim: trim,
|
||
stripBOM: stripBOM
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 49:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
var wrappy = __webpack_require__(11)
|
||
module.exports = wrappy(once)
|
||
module.exports.strict = wrappy(onceStrict)
|
||
|
||
once.proto = once(function () {
|
||
Object.defineProperty(Function.prototype, 'once', {
|
||
value: function () {
|
||
return once(this)
|
||
},
|
||
configurable: true
|
||
})
|
||
|
||
Object.defineProperty(Function.prototype, 'onceStrict', {
|
||
value: function () {
|
||
return onceStrict(this)
|
||
},
|
||
configurable: true
|
||
})
|
||
})
|
||
|
||
function once (fn) {
|
||
var f = function () {
|
||
if (f.called) return f.value
|
||
f.called = true
|
||
return f.value = fn.apply(this, arguments)
|
||
}
|
||
f.called = false
|
||
return f
|
||
}
|
||
|
||
function onceStrict (fn) {
|
||
var f = function () {
|
||
if (f.called)
|
||
throw new Error(f.onceError)
|
||
f.called = true
|
||
return f.value = fn.apply(this, arguments)
|
||
}
|
||
var name = fn.name || 'Function wrapped with `once`'
|
||
f.onceError = name + " shouldn't be called more than once"
|
||
f.called = false
|
||
return f
|
||
}
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 53:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
module.exports = __webpack_require__(352);
|
||
|
||
/***/ }),
|
||
|
||
/***/ 81:
|
||
/***/ (function(module, exports, __webpack_require__) {
|
||
|
||
/**
|
||
* Module dependencies.
|
||
*/
|
||
|
||
const tty = __webpack_require__(867);
|
||
const util = __webpack_require__(669);
|
||
|
||
/**
|
||
* This is the Node.js implementation of `debug()`.
|
||
*/
|
||
|
||
exports.init = init;
|
||
exports.log = log;
|
||
exports.formatArgs = formatArgs;
|
||
exports.save = save;
|
||
exports.load = load;
|
||
exports.useColors = useColors;
|
||
|
||
/**
|
||
* Colors.
|
||
*/
|
||
|
||
exports.colors = [6, 2, 3, 4, 5, 1];
|
||
|
||
try {
|
||
// Optional dependency (as in, doesn't need to be installed, NOT like optionalDependencies in package.json)
|
||
// eslint-disable-next-line import/no-extraneous-dependencies
|
||
const supportsColor = __webpack_require__(247);
|
||
|
||
if (supportsColor && (supportsColor.stderr || supportsColor).level >= 2) {
|
||
exports.colors = [
|
||
20,
|
||
21,
|
||
26,
|
||
27,
|
||
32,
|
||
33,
|
||
38,
|
||
39,
|
||
40,
|
||
41,
|
||
42,
|
||
43,
|
||
44,
|
||
45,
|
||
56,
|
||
57,
|
||
62,
|
||
63,
|
||
68,
|
||
69,
|
||
74,
|
||
75,
|
||
76,
|
||
77,
|
||
78,
|
||
79,
|
||
80,
|
||
81,
|
||
92,
|
||
93,
|
||
98,
|
||
99,
|
||
112,
|
||
113,
|
||
128,
|
||
129,
|
||
134,
|
||
135,
|
||
148,
|
||
149,
|
||
160,
|
||
161,
|
||
162,
|
||
163,
|
||
164,
|
||
165,
|
||
166,
|
||
167,
|
||
168,
|
||
169,
|
||
170,
|
||
171,
|
||
172,
|
||
173,
|
||
178,
|
||
179,
|
||
184,
|
||
185,
|
||
196,
|
||
197,
|
||
198,
|
||
199,
|
||
200,
|
||
201,
|
||
202,
|
||
203,
|
||
204,
|
||
205,
|
||
206,
|
||
207,
|
||
208,
|
||
209,
|
||
214,
|
||
215,
|
||
220,
|
||
221
|
||
];
|
||
}
|
||
} catch (error) {
|
||
// Swallow - we only care if `supports-color` is available; it doesn't have to be.
|
||
}
|
||
|
||
/**
|
||
* Build up the default `inspectOpts` object from the environment variables.
|
||
*
|
||
* $ DEBUG_COLORS=no DEBUG_DEPTH=10 DEBUG_SHOW_HIDDEN=enabled node script.js
|
||
*/
|
||
|
||
exports.inspectOpts = Object.keys(process.env).filter(key => {
|
||
return /^debug_/i.test(key);
|
||
}).reduce((obj, key) => {
|
||
// Camel-case
|
||
const prop = key
|
||
.substring(6)
|
||
.toLowerCase()
|
||
.replace(/_([a-z])/g, (_, k) => {
|
||
return k.toUpperCase();
|
||
});
|
||
|
||
// Coerce string value into JS value
|
||
let val = process.env[key];
|
||
if (/^(yes|on|true|enabled)$/i.test(val)) {
|
||
val = true;
|
||
} else if (/^(no|off|false|disabled)$/i.test(val)) {
|
||
val = false;
|
||
} else if (val === 'null') {
|
||
val = null;
|
||
} else {
|
||
val = Number(val);
|
||
}
|
||
|
||
obj[prop] = val;
|
||
return obj;
|
||
}, {});
|
||
|
||
/**
|
||
* Is stdout a TTY? Colored output is enabled when `true`.
|
||
*/
|
||
|
||
function useColors() {
|
||
return 'colors' in exports.inspectOpts ?
|
||
Boolean(exports.inspectOpts.colors) :
|
||
tty.isatty(process.stderr.fd);
|
||
}
|
||
|
||
/**
|
||
* Adds ANSI color escape codes if enabled.
|
||
*
|
||
* @api public
|
||
*/
|
||
|
||
function formatArgs(args) {
|
||
const {namespace: name, useColors} = this;
|
||
|
||
if (useColors) {
|
||
const c = this.color;
|
||
const colorCode = '\u001B[3' + (c < 8 ? c : '8;5;' + c);
|
||
const prefix = ` ${colorCode};1m${name} \u001B[0m`;
|
||
|
||
args[0] = prefix + args[0].split('\n').join('\n' + prefix);
|
||
args.push(colorCode + 'm+' + module.exports.humanize(this.diff) + '\u001B[0m');
|
||
} else {
|
||
args[0] = getDate() + name + ' ' + args[0];
|
||
}
|
||
}
|
||
|
||
function getDate() {
|
||
if (exports.inspectOpts.hideDate) {
|
||
return '';
|
||
}
|
||
return new Date().toISOString() + ' ';
|
||
}
|
||
|
||
/**
|
||
* Invokes `util.format()` with the specified arguments and writes to stderr.
|
||
*/
|
||
|
||
function log(...args) {
|
||
return process.stderr.write(util.format(...args) + '\n');
|
||
}
|
||
|
||
/**
|
||
* Save `namespaces`.
|
||
*
|
||
* @param {String} namespaces
|
||
* @api private
|
||
*/
|
||
function save(namespaces) {
|
||
if (namespaces) {
|
||
process.env.DEBUG = namespaces;
|
||
} else {
|
||
// If you set a process.env field to null or undefined, it gets cast to the
|
||
// string 'null' or 'undefined'. Just delete instead.
|
||
delete process.env.DEBUG;
|
||
}
|
||
}
|
||
|
||
/**
|
||
* Load `namespaces`.
|
||
*
|
||
* @return {String} returns the previously persisted debug modes
|
||
* @api private
|
||
*/
|
||
|
||
function load() {
|
||
return process.env.DEBUG;
|
||
}
|
||
|
||
/**
|
||
* Init logic for `debug` instances.
|
||
*
|
||
* Create a new `inspectOpts` object in case `useColors` is set
|
||
* differently for a particular `debug` instance.
|
||
*/
|
||
|
||
function init(debug) {
|
||
debug.inspectOpts = {};
|
||
|
||
const keys = Object.keys(exports.inspectOpts);
|
||
for (let i = 0; i < keys.length; i++) {
|
||
debug.inspectOpts[keys[i]] = exports.inspectOpts[keys[i]];
|
||
}
|
||
}
|
||
|
||
module.exports = __webpack_require__(486)(exports);
|
||
|
||
const {formatters} = module.exports;
|
||
|
||
/**
|
||
* Map %o to `util.inspect()`, all on a single line.
|
||
*/
|
||
|
||
formatters.o = function (v) {
|
||
this.inspectOpts.colors = this.useColors;
|
||
return util.inspect(v, this.inspectOpts)
|
||
.replace(/\s*\n\s*/g, ' ');
|
||
};
|
||
|
||
/**
|
||
* Map %O to `util.inspect()`, allowing multiple lines if needed.
|
||
*/
|
||
|
||
formatters.O = function (v) {
|
||
this.inspectOpts.colors = this.useColors;
|
||
return util.inspect(v, this.inspectOpts);
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 82:
|
||
/***/ (function(__unusedmodule, exports) {
|
||
|
||
"use strict";
|
||
|
||
// We use any as a valid input type
|
||
/* eslint-disable @typescript-eslint/no-explicit-any */
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
/**
|
||
* Sanitizes an input into a string so it can be passed into issueCommand safely
|
||
* @param input input to sanitize into a string
|
||
*/
|
||
function toCommandValue(input) {
|
||
if (input === null || input === undefined) {
|
||
return '';
|
||
}
|
||
else if (typeof input === 'string' || input instanceof String) {
|
||
return input;
|
||
}
|
||
return JSON.stringify(input);
|
||
}
|
||
exports.toCommandValue = toCommandValue;
|
||
//# sourceMappingURL=utils.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 87:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("os");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 93:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
module.exports = minimatch
|
||
minimatch.Minimatch = Minimatch
|
||
|
||
var path = { sep: '/' }
|
||
try {
|
||
path = __webpack_require__(622)
|
||
} catch (er) {}
|
||
|
||
var GLOBSTAR = minimatch.GLOBSTAR = Minimatch.GLOBSTAR = {}
|
||
var expand = __webpack_require__(306)
|
||
|
||
var plTypes = {
|
||
'!': { open: '(?:(?!(?:', close: '))[^/]*?)'},
|
||
'?': { open: '(?:', close: ')?' },
|
||
'+': { open: '(?:', close: ')+' },
|
||
'*': { open: '(?:', close: ')*' },
|
||
'@': { open: '(?:', close: ')' }
|
||
}
|
||
|
||
// any single thing other than /
|
||
// don't need to escape / when using new RegExp()
|
||
var qmark = '[^/]'
|
||
|
||
// * => any number of characters
|
||
var star = qmark + '*?'
|
||
|
||
// ** when dots are allowed. Anything goes, except .. and .
|
||
// not (^ or / followed by one or two dots followed by $ or /),
|
||
// followed by anything, any number of times.
|
||
var twoStarDot = '(?:(?!(?:\\\/|^)(?:\\.{1,2})($|\\\/)).)*?'
|
||
|
||
// not a ^ or / followed by a dot,
|
||
// followed by anything, any number of times.
|
||
var twoStarNoDot = '(?:(?!(?:\\\/|^)\\.).)*?'
|
||
|
||
// characters that need to be escaped in RegExp.
|
||
var reSpecials = charSet('().*{}+?[]^$\\!')
|
||
|
||
// "abc" -> { a:true, b:true, c:true }
|
||
function charSet (s) {
|
||
return s.split('').reduce(function (set, c) {
|
||
set[c] = true
|
||
return set
|
||
}, {})
|
||
}
|
||
|
||
// normalizes slashes.
|
||
var slashSplit = /\/+/
|
||
|
||
minimatch.filter = filter
|
||
function filter (pattern, options) {
|
||
options = options || {}
|
||
return function (p, i, list) {
|
||
return minimatch(p, pattern, options)
|
||
}
|
||
}
|
||
|
||
function ext (a, b) {
|
||
a = a || {}
|
||
b = b || {}
|
||
var t = {}
|
||
Object.keys(b).forEach(function (k) {
|
||
t[k] = b[k]
|
||
})
|
||
Object.keys(a).forEach(function (k) {
|
||
t[k] = a[k]
|
||
})
|
||
return t
|
||
}
|
||
|
||
minimatch.defaults = function (def) {
|
||
if (!def || !Object.keys(def).length) return minimatch
|
||
|
||
var orig = minimatch
|
||
|
||
var m = function minimatch (p, pattern, options) {
|
||
return orig.minimatch(p, pattern, ext(def, options))
|
||
}
|
||
|
||
m.Minimatch = function Minimatch (pattern, options) {
|
||
return new orig.Minimatch(pattern, ext(def, options))
|
||
}
|
||
|
||
return m
|
||
}
|
||
|
||
Minimatch.defaults = function (def) {
|
||
if (!def || !Object.keys(def).length) return Minimatch
|
||
return minimatch.defaults(def).Minimatch
|
||
}
|
||
|
||
function minimatch (p, pattern, options) {
|
||
if (typeof pattern !== 'string') {
|
||
throw new TypeError('glob pattern string required')
|
||
}
|
||
|
||
if (!options) options = {}
|
||
|
||
// shortcut: comments match nothing.
|
||
if (!options.nocomment && pattern.charAt(0) === '#') {
|
||
return false
|
||
}
|
||
|
||
// "" only matches ""
|
||
if (pattern.trim() === '') return p === ''
|
||
|
||
return new Minimatch(pattern, options).match(p)
|
||
}
|
||
|
||
function Minimatch (pattern, options) {
|
||
if (!(this instanceof Minimatch)) {
|
||
return new Minimatch(pattern, options)
|
||
}
|
||
|
||
if (typeof pattern !== 'string') {
|
||
throw new TypeError('glob pattern string required')
|
||
}
|
||
|
||
if (!options) options = {}
|
||
pattern = pattern.trim()
|
||
|
||
// windows support: need to use /, not \
|
||
if (path.sep !== '/') {
|
||
pattern = pattern.split(path.sep).join('/')
|
||
}
|
||
|
||
this.options = options
|
||
this.set = []
|
||
this.pattern = pattern
|
||
this.regexp = null
|
||
this.negate = false
|
||
this.comment = false
|
||
this.empty = false
|
||
|
||
// make the set of regexps etc.
|
||
this.make()
|
||
}
|
||
|
||
Minimatch.prototype.debug = function () {}
|
||
|
||
Minimatch.prototype.make = make
|
||
function make () {
|
||
// don't do it more than once.
|
||
if (this._made) return
|
||
|
||
var pattern = this.pattern
|
||
var options = this.options
|
||
|
||
// empty patterns and comments match nothing.
|
||
if (!options.nocomment && pattern.charAt(0) === '#') {
|
||
this.comment = true
|
||
return
|
||
}
|
||
if (!pattern) {
|
||
this.empty = true
|
||
return
|
||
}
|
||
|
||
// step 1: figure out negation, etc.
|
||
this.parseNegate()
|
||
|
||
// step 2: expand braces
|
||
var set = this.globSet = this.braceExpand()
|
||
|
||
if (options.debug) this.debug = console.error
|
||
|
||
this.debug(this.pattern, set)
|
||
|
||
// step 3: now we have a set, so turn each one into a series of path-portion
|
||
// matching patterns.
|
||
// These will be regexps, except in the case of "**", which is
|
||
// set to the GLOBSTAR object for globstar behavior,
|
||
// and will not contain any / characters
|
||
set = this.globParts = set.map(function (s) {
|
||
return s.split(slashSplit)
|
||
})
|
||
|
||
this.debug(this.pattern, set)
|
||
|
||
// glob --> regexps
|
||
set = set.map(function (s, si, set) {
|
||
return s.map(this.parse, this)
|
||
}, this)
|
||
|
||
this.debug(this.pattern, set)
|
||
|
||
// filter out everything that didn't compile properly.
|
||
set = set.filter(function (s) {
|
||
return s.indexOf(false) === -1
|
||
})
|
||
|
||
this.debug(this.pattern, set)
|
||
|
||
this.set = set
|
||
}
|
||
|
||
Minimatch.prototype.parseNegate = parseNegate
|
||
function parseNegate () {
|
||
var pattern = this.pattern
|
||
var negate = false
|
||
var options = this.options
|
||
var negateOffset = 0
|
||
|
||
if (options.nonegate) return
|
||
|
||
for (var i = 0, l = pattern.length
|
||
; i < l && pattern.charAt(i) === '!'
|
||
; i++) {
|
||
negate = !negate
|
||
negateOffset++
|
||
}
|
||
|
||
if (negateOffset) this.pattern = pattern.substr(negateOffset)
|
||
this.negate = negate
|
||
}
|
||
|
||
// Brace expansion:
|
||
// a{b,c}d -> abd acd
|
||
// a{b,}c -> abc ac
|
||
// a{0..3}d -> a0d a1d a2d a3d
|
||
// a{b,c{d,e}f}g -> abg acdfg acefg
|
||
// a{b,c}d{e,f}g -> abdeg acdeg abdeg abdfg
|
||
//
|
||
// Invalid sets are not expanded.
|
||
// a{2..}b -> a{2..}b
|
||
// a{b}c -> a{b}c
|
||
minimatch.braceExpand = function (pattern, options) {
|
||
return braceExpand(pattern, options)
|
||
}
|
||
|
||
Minimatch.prototype.braceExpand = braceExpand
|
||
|
||
function braceExpand (pattern, options) {
|
||
if (!options) {
|
||
if (this instanceof Minimatch) {
|
||
options = this.options
|
||
} else {
|
||
options = {}
|
||
}
|
||
}
|
||
|
||
pattern = typeof pattern === 'undefined'
|
||
? this.pattern : pattern
|
||
|
||
if (typeof pattern === 'undefined') {
|
||
throw new TypeError('undefined pattern')
|
||
}
|
||
|
||
if (options.nobrace ||
|
||
!pattern.match(/\{.*\}/)) {
|
||
// shortcut. no need to expand.
|
||
return [pattern]
|
||
}
|
||
|
||
return expand(pattern)
|
||
}
|
||
|
||
// parse a component of the expanded set.
|
||
// At this point, no pattern may contain "/" in it
|
||
// so we're going to return a 2d array, where each entry is the full
|
||
// pattern, split on '/', and then turned into a regular expression.
|
||
// A regexp is made at the end which joins each array with an
|
||
// escaped /, and another full one which joins each regexp with |.
|
||
//
|
||
// Following the lead of Bash 4.1, note that "**" only has special meaning
|
||
// when it is the *only* thing in a path portion. Otherwise, any series
|
||
// of * is equivalent to a single *. Globstar behavior is enabled by
|
||
// default, and can be disabled by setting options.noglobstar.
|
||
Minimatch.prototype.parse = parse
|
||
var SUBPARSE = {}
|
||
function parse (pattern, isSub) {
|
||
if (pattern.length > 1024 * 64) {
|
||
throw new TypeError('pattern is too long')
|
||
}
|
||
|
||
var options = this.options
|
||
|
||
// shortcuts
|
||
if (!options.noglobstar && pattern === '**') return GLOBSTAR
|
||
if (pattern === '') return ''
|
||
|
||
var re = ''
|
||
var hasMagic = !!options.nocase
|
||
var escaping = false
|
||
// ? => one single character
|
||
var patternListStack = []
|
||
var negativeLists = []
|
||
var stateChar
|
||
var inClass = false
|
||
var reClassStart = -1
|
||
var classStart = -1
|
||
// . and .. never match anything that doesn't start with .,
|
||
// even when options.dot is set.
|
||
var patternStart = pattern.charAt(0) === '.' ? '' // anything
|
||
// not (start or / followed by . or .. followed by / or end)
|
||
: options.dot ? '(?!(?:^|\\\/)\\.{1,2}(?:$|\\\/))'
|
||
: '(?!\\.)'
|
||
var self = this
|
||
|
||
function clearStateChar () {
|
||
if (stateChar) {
|
||
// we had some state-tracking character
|
||
// that wasn't consumed by this pass.
|
||
switch (stateChar) {
|
||
case '*':
|
||
re += star
|
||
hasMagic = true
|
||
break
|
||
case '?':
|
||
re += qmark
|
||
hasMagic = true
|
||
break
|
||
default:
|
||
re += '\\' + stateChar
|
||
break
|
||
}
|
||
self.debug('clearStateChar %j %j', stateChar, re)
|
||
stateChar = false
|
||
}
|
||
}
|
||
|
||
for (var i = 0, len = pattern.length, c
|
||
; (i < len) && (c = pattern.charAt(i))
|
||
; i++) {
|
||
this.debug('%s\t%s %s %j', pattern, i, re, c)
|
||
|
||
// skip over any that are escaped.
|
||
if (escaping && reSpecials[c]) {
|
||
re += '\\' + c
|
||
escaping = false
|
||
continue
|
||
}
|
||
|
||
switch (c) {
|
||
case '/':
|
||
// completely not allowed, even escaped.
|
||
// Should already be path-split by now.
|
||
return false
|
||
|
||
case '\\':
|
||
clearStateChar()
|
||
escaping = true
|
||
continue
|
||
|
||
// the various stateChar values
|
||
// for the "extglob" stuff.
|
||
case '?':
|
||
case '*':
|
||
case '+':
|
||
case '@':
|
||
case '!':
|
||
this.debug('%s\t%s %s %j <-- stateChar', pattern, i, re, c)
|
||
|
||
// all of those are literals inside a class, except that
|
||
// the glob [!a] means [^a] in regexp
|
||
if (inClass) {
|
||
this.debug(' in class')
|
||
if (c === '!' && i === classStart + 1) c = '^'
|
||
re += c
|
||
continue
|
||
}
|
||
|
||
// if we already have a stateChar, then it means
|
||
// that there was something like ** or +? in there.
|
||
// Handle the stateChar, then proceed with this one.
|
||
self.debug('call clearStateChar %j', stateChar)
|
||
clearStateChar()
|
||
stateChar = c
|
||
// if extglob is disabled, then +(asdf|foo) isn't a thing.
|
||
// just clear the statechar *now*, rather than even diving into
|
||
// the patternList stuff.
|
||
if (options.noext) clearStateChar()
|
||
continue
|
||
|
||
case '(':
|
||
if (inClass) {
|
||
re += '('
|
||
continue
|
||
}
|
||
|
||
if (!stateChar) {
|
||
re += '\\('
|
||
continue
|
||
}
|
||
|
||
patternListStack.push({
|
||
type: stateChar,
|
||
start: i - 1,
|
||
reStart: re.length,
|
||
open: plTypes[stateChar].open,
|
||
close: plTypes[stateChar].close
|
||
})
|
||
// negation is (?:(?!js)[^/]*)
|
||
re += stateChar === '!' ? '(?:(?!(?:' : '(?:'
|
||
this.debug('plType %j %j', stateChar, re)
|
||
stateChar = false
|
||
continue
|
||
|
||
case ')':
|
||
if (inClass || !patternListStack.length) {
|
||
re += '\\)'
|
||
continue
|
||
}
|
||
|
||
clearStateChar()
|
||
hasMagic = true
|
||
var pl = patternListStack.pop()
|
||
// negation is (?:(?!js)[^/]*)
|
||
// The others are (?:<pattern>)<type>
|
||
re += pl.close
|
||
if (pl.type === '!') {
|
||
negativeLists.push(pl)
|
||
}
|
||
pl.reEnd = re.length
|
||
continue
|
||
|
||
case '|':
|
||
if (inClass || !patternListStack.length || escaping) {
|
||
re += '\\|'
|
||
escaping = false
|
||
continue
|
||
}
|
||
|
||
clearStateChar()
|
||
re += '|'
|
||
continue
|
||
|
||
// these are mostly the same in regexp and glob
|
||
case '[':
|
||
// swallow any state-tracking char before the [
|
||
clearStateChar()
|
||
|
||
if (inClass) {
|
||
re += '\\' + c
|
||
continue
|
||
}
|
||
|
||
inClass = true
|
||
classStart = i
|
||
reClassStart = re.length
|
||
re += c
|
||
continue
|
||
|
||
case ']':
|
||
// a right bracket shall lose its special
|
||
// meaning and represent itself in
|
||
// a bracket expression if it occurs
|
||
// first in the list. -- POSIX.2 2.8.3.2
|
||
if (i === classStart + 1 || !inClass) {
|
||
re += '\\' + c
|
||
escaping = false
|
||
continue
|
||
}
|
||
|
||
// handle the case where we left a class open.
|
||
// "[z-a]" is valid, equivalent to "\[z-a\]"
|
||
if (inClass) {
|
||
// split where the last [ was, make sure we don't have
|
||
// an invalid re. if so, re-walk the contents of the
|
||
// would-be class to re-translate any characters that
|
||
// were passed through as-is
|
||
// TODO: It would probably be faster to determine this
|
||
// without a try/catch and a new RegExp, but it's tricky
|
||
// to do safely. For now, this is safe and works.
|
||
var cs = pattern.substring(classStart + 1, i)
|
||
try {
|
||
RegExp('[' + cs + ']')
|
||
} catch (er) {
|
||
// not a valid class!
|
||
var sp = this.parse(cs, SUBPARSE)
|
||
re = re.substr(0, reClassStart) + '\\[' + sp[0] + '\\]'
|
||
hasMagic = hasMagic || sp[1]
|
||
inClass = false
|
||
continue
|
||
}
|
||
}
|
||
|
||
// finish up the class.
|
||
hasMagic = true
|
||
inClass = false
|
||
re += c
|
||
continue
|
||
|
||
default:
|
||
// swallow any state char that wasn't consumed
|
||
clearStateChar()
|
||
|
||
if (escaping) {
|
||
// no need
|
||
escaping = false
|
||
} else if (reSpecials[c]
|
||
&& !(c === '^' && inClass)) {
|
||
re += '\\'
|
||
}
|
||
|
||
re += c
|
||
|
||
} // switch
|
||
} // for
|
||
|
||
// handle the case where we left a class open.
|
||
// "[abc" is valid, equivalent to "\[abc"
|
||
if (inClass) {
|
||
// split where the last [ was, and escape it
|
||
// this is a huge pita. We now have to re-walk
|
||
// the contents of the would-be class to re-translate
|
||
// any characters that were passed through as-is
|
||
cs = pattern.substr(classStart + 1)
|
||
sp = this.parse(cs, SUBPARSE)
|
||
re = re.substr(0, reClassStart) + '\\[' + sp[0]
|
||
hasMagic = hasMagic || sp[1]
|
||
}
|
||
|
||
// handle the case where we had a +( thing at the *end*
|
||
// of the pattern.
|
||
// each pattern list stack adds 3 chars, and we need to go through
|
||
// and escape any | chars that were passed through as-is for the regexp.
|
||
// Go through and escape them, taking care not to double-escape any
|
||
// | chars that were already escaped.
|
||
for (pl = patternListStack.pop(); pl; pl = patternListStack.pop()) {
|
||
var tail = re.slice(pl.reStart + pl.open.length)
|
||
this.debug('setting tail', re, pl)
|
||
// maybe some even number of \, then maybe 1 \, followed by a |
|
||
tail = tail.replace(/((?:\\{2}){0,64})(\\?)\|/g, function (_, $1, $2) {
|
||
if (!$2) {
|
||
// the | isn't already escaped, so escape it.
|
||
$2 = '\\'
|
||
}
|
||
|
||
// need to escape all those slashes *again*, without escaping the
|
||
// one that we need for escaping the | character. As it works out,
|
||
// escaping an even number of slashes can be done by simply repeating
|
||
// it exactly after itself. That's why this trick works.
|
||
//
|
||
// I am sorry that you have to see this.
|
||
return $1 + $1 + $2 + '|'
|
||
})
|
||
|
||
this.debug('tail=%j\n %s', tail, tail, pl, re)
|
||
var t = pl.type === '*' ? star
|
||
: pl.type === '?' ? qmark
|
||
: '\\' + pl.type
|
||
|
||
hasMagic = true
|
||
re = re.slice(0, pl.reStart) + t + '\\(' + tail
|
||
}
|
||
|
||
// handle trailing things that only matter at the very end.
|
||
clearStateChar()
|
||
if (escaping) {
|
||
// trailing \\
|
||
re += '\\\\'
|
||
}
|
||
|
||
// only need to apply the nodot start if the re starts with
|
||
// something that could conceivably capture a dot
|
||
var addPatternStart = false
|
||
switch (re.charAt(0)) {
|
||
case '.':
|
||
case '[':
|
||
case '(': addPatternStart = true
|
||
}
|
||
|
||
// Hack to work around lack of negative lookbehind in JS
|
||
// A pattern like: *.!(x).!(y|z) needs to ensure that a name
|
||
// like 'a.xyz.yz' doesn't match. So, the first negative
|
||
// lookahead, has to look ALL the way ahead, to the end of
|
||
// the pattern.
|
||
for (var n = negativeLists.length - 1; n > -1; n--) {
|
||
var nl = negativeLists[n]
|
||
|
||
var nlBefore = re.slice(0, nl.reStart)
|
||
var nlFirst = re.slice(nl.reStart, nl.reEnd - 8)
|
||
var nlLast = re.slice(nl.reEnd - 8, nl.reEnd)
|
||
var nlAfter = re.slice(nl.reEnd)
|
||
|
||
nlLast += nlAfter
|
||
|
||
// Handle nested stuff like *(*.js|!(*.json)), where open parens
|
||
// mean that we should *not* include the ) in the bit that is considered
|
||
// "after" the negated section.
|
||
var openParensBefore = nlBefore.split('(').length - 1
|
||
var cleanAfter = nlAfter
|
||
for (i = 0; i < openParensBefore; i++) {
|
||
cleanAfter = cleanAfter.replace(/\)[+*?]?/, '')
|
||
}
|
||
nlAfter = cleanAfter
|
||
|
||
var dollar = ''
|
||
if (nlAfter === '' && isSub !== SUBPARSE) {
|
||
dollar = '$'
|
||
}
|
||
var newRe = nlBefore + nlFirst + nlAfter + dollar + nlLast
|
||
re = newRe
|
||
}
|
||
|
||
// if the re is not "" at this point, then we need to make sure
|
||
// it doesn't match against an empty path part.
|
||
// Otherwise a/* will match a/, which it should not.
|
||
if (re !== '' && hasMagic) {
|
||
re = '(?=.)' + re
|
||
}
|
||
|
||
if (addPatternStart) {
|
||
re = patternStart + re
|
||
}
|
||
|
||
// parsing just a piece of a larger pattern.
|
||
if (isSub === SUBPARSE) {
|
||
return [re, hasMagic]
|
||
}
|
||
|
||
// skip the regexp for non-magical patterns
|
||
// unescape anything in it, though, so that it'll be
|
||
// an exact match against a file etc.
|
||
if (!hasMagic) {
|
||
return globUnescape(pattern)
|
||
}
|
||
|
||
var flags = options.nocase ? 'i' : ''
|
||
try {
|
||
var regExp = new RegExp('^' + re + '$', flags)
|
||
} catch (er) {
|
||
// If it was an invalid regular expression, then it can't match
|
||
// anything. This trick looks for a character after the end of
|
||
// the string, which is of course impossible, except in multi-line
|
||
// mode, but it's not a /m regex.
|
||
return new RegExp('$.')
|
||
}
|
||
|
||
regExp._glob = pattern
|
||
regExp._src = re
|
||
|
||
return regExp
|
||
}
|
||
|
||
minimatch.makeRe = function (pattern, options) {
|
||
return new Minimatch(pattern, options || {}).makeRe()
|
||
}
|
||
|
||
Minimatch.prototype.makeRe = makeRe
|
||
function makeRe () {
|
||
if (this.regexp || this.regexp === false) return this.regexp
|
||
|
||
// at this point, this.set is a 2d array of partial
|
||
// pattern strings, or "**".
|
||
//
|
||
// It's better to use .match(). This function shouldn't
|
||
// be used, really, but it's pretty convenient sometimes,
|
||
// when you just want to work with a regex.
|
||
var set = this.set
|
||
|
||
if (!set.length) {
|
||
this.regexp = false
|
||
return this.regexp
|
||
}
|
||
var options = this.options
|
||
|
||
var twoStar = options.noglobstar ? star
|
||
: options.dot ? twoStarDot
|
||
: twoStarNoDot
|
||
var flags = options.nocase ? 'i' : ''
|
||
|
||
var re = set.map(function (pattern) {
|
||
return pattern.map(function (p) {
|
||
return (p === GLOBSTAR) ? twoStar
|
||
: (typeof p === 'string') ? regExpEscape(p)
|
||
: p._src
|
||
}).join('\\\/')
|
||
}).join('|')
|
||
|
||
// must match entire pattern
|
||
// ending in a * or ** will make it less strict.
|
||
re = '^(?:' + re + ')$'
|
||
|
||
// can match anything, as long as it's not this.
|
||
if (this.negate) re = '^(?!' + re + ').*$'
|
||
|
||
try {
|
||
this.regexp = new RegExp(re, flags)
|
||
} catch (ex) {
|
||
this.regexp = false
|
||
}
|
||
return this.regexp
|
||
}
|
||
|
||
minimatch.match = function (list, pattern, options) {
|
||
options = options || {}
|
||
var mm = new Minimatch(pattern, options)
|
||
list = list.filter(function (f) {
|
||
return mm.match(f)
|
||
})
|
||
if (mm.options.nonull && !list.length) {
|
||
list.push(pattern)
|
||
}
|
||
return list
|
||
}
|
||
|
||
Minimatch.prototype.match = match
|
||
function match (f, partial) {
|
||
this.debug('match', f, this.pattern)
|
||
// short-circuit in the case of busted things.
|
||
// comments, etc.
|
||
if (this.comment) return false
|
||
if (this.empty) return f === ''
|
||
|
||
if (f === '/' && partial) return true
|
||
|
||
var options = this.options
|
||
|
||
// windows: need to use /, not \
|
||
if (path.sep !== '/') {
|
||
f = f.split(path.sep).join('/')
|
||
}
|
||
|
||
// treat the test path as a set of pathparts.
|
||
f = f.split(slashSplit)
|
||
this.debug(this.pattern, 'split', f)
|
||
|
||
// just ONE of the pattern sets in this.set needs to match
|
||
// in order for it to be valid. If negating, then just one
|
||
// match means that we have failed.
|
||
// Either way, return on the first hit.
|
||
|
||
var set = this.set
|
||
this.debug(this.pattern, 'set', set)
|
||
|
||
// Find the basename of the path by looking for the last non-empty segment
|
||
var filename
|
||
var i
|
||
for (i = f.length - 1; i >= 0; i--) {
|
||
filename = f[i]
|
||
if (filename) break
|
||
}
|
||
|
||
for (i = 0; i < set.length; i++) {
|
||
var pattern = set[i]
|
||
var file = f
|
||
if (options.matchBase && pattern.length === 1) {
|
||
file = [filename]
|
||
}
|
||
var hit = this.matchOne(file, pattern, partial)
|
||
if (hit) {
|
||
if (options.flipNegate) return true
|
||
return !this.negate
|
||
}
|
||
}
|
||
|
||
// didn't get any hits. this is success if it's a negative
|
||
// pattern, failure otherwise.
|
||
if (options.flipNegate) return false
|
||
return this.negate
|
||
}
|
||
|
||
// set partial to true to test if, for example,
|
||
// "/a/b" matches the start of "/*/b/*/d"
|
||
// Partial means, if you run out of file before you run
|
||
// out of pattern, then that's fine, as long as all
|
||
// the parts match.
|
||
Minimatch.prototype.matchOne = function (file, pattern, partial) {
|
||
var options = this.options
|
||
|
||
this.debug('matchOne',
|
||
{ 'this': this, file: file, pattern: pattern })
|
||
|
||
this.debug('matchOne', file.length, pattern.length)
|
||
|
||
for (var fi = 0,
|
||
pi = 0,
|
||
fl = file.length,
|
||
pl = pattern.length
|
||
; (fi < fl) && (pi < pl)
|
||
; fi++, pi++) {
|
||
this.debug('matchOne loop')
|
||
var p = pattern[pi]
|
||
var f = file[fi]
|
||
|
||
this.debug(pattern, p, f)
|
||
|
||
// should be impossible.
|
||
// some invalid regexp stuff in the set.
|
||
if (p === false) return false
|
||
|
||
if (p === GLOBSTAR) {
|
||
this.debug('GLOBSTAR', [pattern, p, f])
|
||
|
||
// "**"
|
||
// a/**/b/**/c would match the following:
|
||
// a/b/x/y/z/c
|
||
// a/x/y/z/b/c
|
||
// a/b/x/b/x/c
|
||
// a/b/c
|
||
// To do this, take the rest of the pattern after
|
||
// the **, and see if it would match the file remainder.
|
||
// If so, return success.
|
||
// If not, the ** "swallows" a segment, and try again.
|
||
// This is recursively awful.
|
||
//
|
||
// a/**/b/**/c matching a/b/x/y/z/c
|
||
// - a matches a
|
||
// - doublestar
|
||
// - matchOne(b/x/y/z/c, b/**/c)
|
||
// - b matches b
|
||
// - doublestar
|
||
// - matchOne(x/y/z/c, c) -> no
|
||
// - matchOne(y/z/c, c) -> no
|
||
// - matchOne(z/c, c) -> no
|
||
// - matchOne(c, c) yes, hit
|
||
var fr = fi
|
||
var pr = pi + 1
|
||
if (pr === pl) {
|
||
this.debug('** at the end')
|
||
// a ** at the end will just swallow the rest.
|
||
// We have found a match.
|
||
// however, it will not swallow /.x, unless
|
||
// options.dot is set.
|
||
// . and .. are *never* matched by **, for explosively
|
||
// exponential reasons.
|
||
for (; fi < fl; fi++) {
|
||
if (file[fi] === '.' || file[fi] === '..' ||
|
||
(!options.dot && file[fi].charAt(0) === '.')) return false
|
||
}
|
||
return true
|
||
}
|
||
|
||
// ok, let's see if we can swallow whatever we can.
|
||
while (fr < fl) {
|
||
var swallowee = file[fr]
|
||
|
||
this.debug('\nglobstar while', file, fr, pattern, pr, swallowee)
|
||
|
||
// XXX remove this slice. Just pass the start index.
|
||
if (this.matchOne(file.slice(fr), pattern.slice(pr), partial)) {
|
||
this.debug('globstar found match!', fr, fl, swallowee)
|
||
// found a match.
|
||
return true
|
||
} else {
|
||
// can't swallow "." or ".." ever.
|
||
// can only swallow ".foo" when explicitly asked.
|
||
if (swallowee === '.' || swallowee === '..' ||
|
||
(!options.dot && swallowee.charAt(0) === '.')) {
|
||
this.debug('dot detected!', file, fr, pattern, pr)
|
||
break
|
||
}
|
||
|
||
// ** swallows a segment, and continue.
|
||
this.debug('globstar swallow a segment, and continue')
|
||
fr++
|
||
}
|
||
}
|
||
|
||
// no match was found.
|
||
// However, in partial mode, we can't say this is necessarily over.
|
||
// If there's more *pattern* left, then
|
||
if (partial) {
|
||
// ran out of file
|
||
this.debug('\n>>> no match, partial?', file, fr, pattern, pr)
|
||
if (fr === fl) return true
|
||
}
|
||
return false
|
||
}
|
||
|
||
// something other than **
|
||
// non-magic patterns just have to match exactly
|
||
// patterns with magic have been turned into regexps.
|
||
var hit
|
||
if (typeof p === 'string') {
|
||
if (options.nocase) {
|
||
hit = f.toLowerCase() === p.toLowerCase()
|
||
} else {
|
||
hit = f === p
|
||
}
|
||
this.debug('string match', p, f, hit)
|
||
} else {
|
||
hit = f.match(p)
|
||
this.debug('pattern match', p, f, hit)
|
||
}
|
||
|
||
if (!hit) return false
|
||
}
|
||
|
||
// Note: ending in / means that we'll get a final ""
|
||
// at the end of the pattern. This can only match a
|
||
// corresponding "" at the end of the file.
|
||
// If the file ends in /, then it can only match a
|
||
// a pattern that ends in /, unless the pattern just
|
||
// doesn't have any more for it. But, a/b/ should *not*
|
||
// match "a/b/*", even though "" matches against the
|
||
// [^/]*? pattern, except in partial mode, where it might
|
||
// simply not be reached yet.
|
||
// However, a/b/ should still satisfy a/*
|
||
|
||
// now either we fell off the end of the pattern, or we're done.
|
||
if (fi === fl && pi === pl) {
|
||
// ran out of pattern and filename at the same time.
|
||
// an exact hit!
|
||
return true
|
||
} else if (fi === fl) {
|
||
// ran out of file, but still had pattern left.
|
||
// this is ok if we're doing the match as part of
|
||
// a glob fs traversal.
|
||
return partial
|
||
} else if (pi === pl) {
|
||
// ran out of pattern, still have file left.
|
||
// this is only acceptable if we're on the very last
|
||
// empty segment of a file with a trailing slash.
|
||
// a/* should match a/b/
|
||
var emptyFileEnd = (fi === fl - 1) && (file[fi] === '')
|
||
return emptyFileEnd
|
||
}
|
||
|
||
// should be unreachable.
|
||
throw new Error('wtf?')
|
||
}
|
||
|
||
// replace stuff like \* with *
|
||
function globUnescape (s) {
|
||
return s.replace(/\\(.)/g, '$1')
|
||
}
|
||
|
||
function regExpEscape (s) {
|
||
return s.replace(/[-[\]{}()*+?.,\\^$|#\s]/g, '\\$&')
|
||
}
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 102:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
// For internal use, subject to change.
|
||
var __importStar = (this && this.__importStar) || function (mod) {
|
||
if (mod && mod.__esModule) return mod;
|
||
var result = {};
|
||
if (mod != null) for (var k in mod) if (Object.hasOwnProperty.call(mod, k)) result[k] = mod[k];
|
||
result["default"] = mod;
|
||
return result;
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
// We use any as a valid input type
|
||
/* eslint-disable @typescript-eslint/no-explicit-any */
|
||
const fs = __importStar(__webpack_require__(747));
|
||
const os = __importStar(__webpack_require__(87));
|
||
const utils_1 = __webpack_require__(82);
|
||
function issueCommand(command, message) {
|
||
const filePath = process.env[`GITHUB_${command}`];
|
||
if (!filePath) {
|
||
throw new Error(`Unable to find environment variable for file command ${command}`);
|
||
}
|
||
if (!fs.existsSync(filePath)) {
|
||
throw new Error(`Missing file at path: ${filePath}`);
|
||
}
|
||
fs.appendFileSync(filePath, `${utils_1.toCommandValue(message)}${os.EOL}`, {
|
||
encoding: 'utf8'
|
||
});
|
||
}
|
||
exports.issueCommand = issueCommand;
|
||
//# sourceMappingURL=file-command.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 104:
|
||
/***/ (function(module) {
|
||
|
||
"use strict";
|
||
|
||
|
||
/**
|
||
* Determines whether the payload is an error thrown by Axios
|
||
*
|
||
* @param {*} payload The value to test
|
||
* @returns {boolean} True if the payload is an error thrown by Axios, otherwise false
|
||
*/
|
||
module.exports = function isAxiosError(payload) {
|
||
return (typeof payload === 'object') && (payload.isAxiosError === true);
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 106:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var pkg = __webpack_require__(361);
|
||
|
||
var validators = {};
|
||
|
||
// eslint-disable-next-line func-names
|
||
['object', 'boolean', 'number', 'function', 'string', 'symbol'].forEach(function(type, i) {
|
||
validators[type] = function validator(thing) {
|
||
return typeof thing === type || 'a' + (i < 1 ? 'n ' : ' ') + type;
|
||
};
|
||
});
|
||
|
||
var deprecatedWarnings = {};
|
||
var currentVerArr = pkg.version.split('.');
|
||
|
||
/**
|
||
* Compare package versions
|
||
* @param {string} version
|
||
* @param {string?} thanVersion
|
||
* @returns {boolean}
|
||
*/
|
||
function isOlderVersion(version, thanVersion) {
|
||
var pkgVersionArr = thanVersion ? thanVersion.split('.') : currentVerArr;
|
||
var destVer = version.split('.');
|
||
for (var i = 0; i < 3; i++) {
|
||
if (pkgVersionArr[i] > destVer[i]) {
|
||
return true;
|
||
} else if (pkgVersionArr[i] < destVer[i]) {
|
||
return false;
|
||
}
|
||
}
|
||
return false;
|
||
}
|
||
|
||
/**
|
||
* Transitional option validator
|
||
* @param {function|boolean?} validator
|
||
* @param {string?} version
|
||
* @param {string} message
|
||
* @returns {function}
|
||
*/
|
||
validators.transitional = function transitional(validator, version, message) {
|
||
var isDeprecated = version && isOlderVersion(version);
|
||
|
||
function formatMessage(opt, desc) {
|
||
return '[Axios v' + pkg.version + '] Transitional option \'' + opt + '\'' + desc + (message ? '. ' + message : '');
|
||
}
|
||
|
||
// eslint-disable-next-line func-names
|
||
return function(value, opt, opts) {
|
||
if (validator === false) {
|
||
throw new Error(formatMessage(opt, ' has been removed in ' + version));
|
||
}
|
||
|
||
if (isDeprecated && !deprecatedWarnings[opt]) {
|
||
deprecatedWarnings[opt] = true;
|
||
// eslint-disable-next-line no-console
|
||
console.warn(
|
||
formatMessage(
|
||
opt,
|
||
' has been deprecated since v' + version + ' and will be removed in the near future'
|
||
)
|
||
);
|
||
}
|
||
|
||
return validator ? validator(value, opt, opts) : true;
|
||
};
|
||
};
|
||
|
||
/**
|
||
* Assert object's properties type
|
||
* @param {object} options
|
||
* @param {object} schema
|
||
* @param {boolean?} allowUnknown
|
||
*/
|
||
|
||
function assertOptions(options, schema, allowUnknown) {
|
||
if (typeof options !== 'object') {
|
||
throw new TypeError('options must be an object');
|
||
}
|
||
var keys = Object.keys(options);
|
||
var i = keys.length;
|
||
while (i-- > 0) {
|
||
var opt = keys[i];
|
||
var validator = schema[opt];
|
||
if (validator) {
|
||
var value = options[opt];
|
||
var result = value === undefined || validator(value, opt, options);
|
||
if (result !== true) {
|
||
throw new TypeError('option ' + opt + ' must be ' + result);
|
||
}
|
||
continue;
|
||
}
|
||
if (allowUnknown !== true) {
|
||
throw Error('Unknown option ' + opt);
|
||
}
|
||
}
|
||
}
|
||
|
||
module.exports = {
|
||
isOlderVersion: isOlderVersion,
|
||
assertOptions: assertOptions,
|
||
validators: validators
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 117:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
// Copyright Joyent, Inc. and other Node contributors.
|
||
//
|
||
// Permission is hereby granted, free of charge, to any person obtaining a
|
||
// copy of this software and associated documentation files (the
|
||
// "Software"), to deal in the Software without restriction, including
|
||
// without limitation the rights to use, copy, modify, merge, publish,
|
||
// distribute, sublicense, and/or sell copies of the Software, and to permit
|
||
// persons to whom the Software is furnished to do so, subject to the
|
||
// following conditions:
|
||
//
|
||
// The above copyright notice and this permission notice shall be included
|
||
// in all copies or substantial portions of the Software.
|
||
//
|
||
// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
|
||
// OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
||
// MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN
|
||
// NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM,
|
||
// DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
|
||
// OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
|
||
// USE OR OTHER DEALINGS IN THE SOFTWARE.
|
||
|
||
var pathModule = __webpack_require__(622);
|
||
var isWindows = process.platform === 'win32';
|
||
var fs = __webpack_require__(747);
|
||
|
||
// JavaScript implementation of realpath, ported from node pre-v6
|
||
|
||
var DEBUG = process.env.NODE_DEBUG && /fs/.test(process.env.NODE_DEBUG);
|
||
|
||
function rethrow() {
|
||
// Only enable in debug mode. A backtrace uses ~1000 bytes of heap space and
|
||
// is fairly slow to generate.
|
||
var callback;
|
||
if (DEBUG) {
|
||
var backtrace = new Error;
|
||
callback = debugCallback;
|
||
} else
|
||
callback = missingCallback;
|
||
|
||
return callback;
|
||
|
||
function debugCallback(err) {
|
||
if (err) {
|
||
backtrace.message = err.message;
|
||
err = backtrace;
|
||
missingCallback(err);
|
||
}
|
||
}
|
||
|
||
function missingCallback(err) {
|
||
if (err) {
|
||
if (process.throwDeprecation)
|
||
throw err; // Forgot a callback but don't know where? Use NODE_DEBUG=fs
|
||
else if (!process.noDeprecation) {
|
||
var msg = 'fs: missing callback ' + (err.stack || err.message);
|
||
if (process.traceDeprecation)
|
||
console.trace(msg);
|
||
else
|
||
console.error(msg);
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
function maybeCallback(cb) {
|
||
return typeof cb === 'function' ? cb : rethrow();
|
||
}
|
||
|
||
var normalize = pathModule.normalize;
|
||
|
||
// Regexp that finds the next partion of a (partial) path
|
||
// result is [base_with_slash, base], e.g. ['somedir/', 'somedir']
|
||
if (isWindows) {
|
||
var nextPartRe = /(.*?)(?:[\/\\]+|$)/g;
|
||
} else {
|
||
var nextPartRe = /(.*?)(?:[\/]+|$)/g;
|
||
}
|
||
|
||
// Regex to find the device root, including trailing slash. E.g. 'c:\\'.
|
||
if (isWindows) {
|
||
var splitRootRe = /^(?:[a-zA-Z]:|[\\\/]{2}[^\\\/]+[\\\/][^\\\/]+)?[\\\/]*/;
|
||
} else {
|
||
var splitRootRe = /^[\/]*/;
|
||
}
|
||
|
||
exports.realpathSync = function realpathSync(p, cache) {
|
||
// make p is absolute
|
||
p = pathModule.resolve(p);
|
||
|
||
if (cache && Object.prototype.hasOwnProperty.call(cache, p)) {
|
||
return cache[p];
|
||
}
|
||
|
||
var original = p,
|
||
seenLinks = {},
|
||
knownHard = {};
|
||
|
||
// current character position in p
|
||
var pos;
|
||
// the partial path so far, including a trailing slash if any
|
||
var current;
|
||
// the partial path without a trailing slash (except when pointing at a root)
|
||
var base;
|
||
// the partial path scanned in the previous round, with slash
|
||
var previous;
|
||
|
||
start();
|
||
|
||
function start() {
|
||
// Skip over roots
|
||
var m = splitRootRe.exec(p);
|
||
pos = m[0].length;
|
||
current = m[0];
|
||
base = m[0];
|
||
previous = '';
|
||
|
||
// On windows, check that the root exists. On unix there is no need.
|
||
if (isWindows && !knownHard[base]) {
|
||
fs.lstatSync(base);
|
||
knownHard[base] = true;
|
||
}
|
||
}
|
||
|
||
// walk down the path, swapping out linked pathparts for their real
|
||
// values
|
||
// NB: p.length changes.
|
||
while (pos < p.length) {
|
||
// find the next part
|
||
nextPartRe.lastIndex = pos;
|
||
var result = nextPartRe.exec(p);
|
||
previous = current;
|
||
current += result[0];
|
||
base = previous + result[1];
|
||
pos = nextPartRe.lastIndex;
|
||
|
||
// continue if not a symlink
|
||
if (knownHard[base] || (cache && cache[base] === base)) {
|
||
continue;
|
||
}
|
||
|
||
var resolvedLink;
|
||
if (cache && Object.prototype.hasOwnProperty.call(cache, base)) {
|
||
// some known symbolic link. no need to stat again.
|
||
resolvedLink = cache[base];
|
||
} else {
|
||
var stat = fs.lstatSync(base);
|
||
if (!stat.isSymbolicLink()) {
|
||
knownHard[base] = true;
|
||
if (cache) cache[base] = base;
|
||
continue;
|
||
}
|
||
|
||
// read the link if it wasn't read before
|
||
// dev/ino always return 0 on windows, so skip the check.
|
||
var linkTarget = null;
|
||
if (!isWindows) {
|
||
var id = stat.dev.toString(32) + ':' + stat.ino.toString(32);
|
||
if (seenLinks.hasOwnProperty(id)) {
|
||
linkTarget = seenLinks[id];
|
||
}
|
||
}
|
||
if (linkTarget === null) {
|
||
fs.statSync(base);
|
||
linkTarget = fs.readlinkSync(base);
|
||
}
|
||
resolvedLink = pathModule.resolve(previous, linkTarget);
|
||
// track this, if given a cache.
|
||
if (cache) cache[base] = resolvedLink;
|
||
if (!isWindows) seenLinks[id] = linkTarget;
|
||
}
|
||
|
||
// resolve the link, then start over
|
||
p = pathModule.resolve(resolvedLink, p.slice(pos));
|
||
start();
|
||
}
|
||
|
||
if (cache) cache[original] = p;
|
||
|
||
return p;
|
||
};
|
||
|
||
|
||
exports.realpath = function realpath(p, cache, cb) {
|
||
if (typeof cb !== 'function') {
|
||
cb = maybeCallback(cache);
|
||
cache = null;
|
||
}
|
||
|
||
// make p is absolute
|
||
p = pathModule.resolve(p);
|
||
|
||
if (cache && Object.prototype.hasOwnProperty.call(cache, p)) {
|
||
return process.nextTick(cb.bind(null, null, cache[p]));
|
||
}
|
||
|
||
var original = p,
|
||
seenLinks = {},
|
||
knownHard = {};
|
||
|
||
// current character position in p
|
||
var pos;
|
||
// the partial path so far, including a trailing slash if any
|
||
var current;
|
||
// the partial path without a trailing slash (except when pointing at a root)
|
||
var base;
|
||
// the partial path scanned in the previous round, with slash
|
||
var previous;
|
||
|
||
start();
|
||
|
||
function start() {
|
||
// Skip over roots
|
||
var m = splitRootRe.exec(p);
|
||
pos = m[0].length;
|
||
current = m[0];
|
||
base = m[0];
|
||
previous = '';
|
||
|
||
// On windows, check that the root exists. On unix there is no need.
|
||
if (isWindows && !knownHard[base]) {
|
||
fs.lstat(base, function(err) {
|
||
if (err) return cb(err);
|
||
knownHard[base] = true;
|
||
LOOP();
|
||
});
|
||
} else {
|
||
process.nextTick(LOOP);
|
||
}
|
||
}
|
||
|
||
// walk down the path, swapping out linked pathparts for their real
|
||
// values
|
||
function LOOP() {
|
||
// stop if scanned past end of path
|
||
if (pos >= p.length) {
|
||
if (cache) cache[original] = p;
|
||
return cb(null, p);
|
||
}
|
||
|
||
// find the next part
|
||
nextPartRe.lastIndex = pos;
|
||
var result = nextPartRe.exec(p);
|
||
previous = current;
|
||
current += result[0];
|
||
base = previous + result[1];
|
||
pos = nextPartRe.lastIndex;
|
||
|
||
// continue if not a symlink
|
||
if (knownHard[base] || (cache && cache[base] === base)) {
|
||
return process.nextTick(LOOP);
|
||
}
|
||
|
||
if (cache && Object.prototype.hasOwnProperty.call(cache, base)) {
|
||
// known symbolic link. no need to stat again.
|
||
return gotResolvedLink(cache[base]);
|
||
}
|
||
|
||
return fs.lstat(base, gotStat);
|
||
}
|
||
|
||
function gotStat(err, stat) {
|
||
if (err) return cb(err);
|
||
|
||
// if not a symlink, skip to the next path part
|
||
if (!stat.isSymbolicLink()) {
|
||
knownHard[base] = true;
|
||
if (cache) cache[base] = base;
|
||
return process.nextTick(LOOP);
|
||
}
|
||
|
||
// stat & read the link if not read before
|
||
// call gotTarget as soon as the link target is known
|
||
// dev/ino always return 0 on windows, so skip the check.
|
||
if (!isWindows) {
|
||
var id = stat.dev.toString(32) + ':' + stat.ino.toString(32);
|
||
if (seenLinks.hasOwnProperty(id)) {
|
||
return gotTarget(null, seenLinks[id], base);
|
||
}
|
||
}
|
||
fs.stat(base, function(err) {
|
||
if (err) return cb(err);
|
||
|
||
fs.readlink(base, function(err, target) {
|
||
if (!isWindows) seenLinks[id] = target;
|
||
gotTarget(err, target);
|
||
});
|
||
});
|
||
}
|
||
|
||
function gotTarget(err, target, base) {
|
||
if (err) return cb(err);
|
||
|
||
var resolvedLink = pathModule.resolve(previous, target);
|
||
if (cache) cache[base] = resolvedLink;
|
||
gotResolvedLink(resolvedLink);
|
||
}
|
||
|
||
function gotResolvedLink(resolvedLink) {
|
||
// resolve the link, then start over
|
||
p = pathModule.resolve(resolvedLink, p.slice(pos));
|
||
start();
|
||
}
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 133:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var utils = __webpack_require__(35);
|
||
|
||
function encode(val) {
|
||
return encodeURIComponent(val).
|
||
replace(/%3A/gi, ':').
|
||
replace(/%24/g, '$').
|
||
replace(/%2C/gi, ',').
|
||
replace(/%20/g, '+').
|
||
replace(/%5B/gi, '[').
|
||
replace(/%5D/gi, ']');
|
||
}
|
||
|
||
/**
|
||
* Build a URL by appending params to the end
|
||
*
|
||
* @param {string} url The base of the url (e.g., http://www.google.com)
|
||
* @param {object} [params] The params to be appended
|
||
* @returns {string} The formatted url
|
||
*/
|
||
module.exports = function buildURL(url, params, paramsSerializer) {
|
||
/*eslint no-param-reassign:0*/
|
||
if (!params) {
|
||
return url;
|
||
}
|
||
|
||
var serializedParams;
|
||
if (paramsSerializer) {
|
||
serializedParams = paramsSerializer(params);
|
||
} else if (utils.isURLSearchParams(params)) {
|
||
serializedParams = params.toString();
|
||
} else {
|
||
var parts = [];
|
||
|
||
utils.forEach(params, function serialize(val, key) {
|
||
if (val === null || typeof val === 'undefined') {
|
||
return;
|
||
}
|
||
|
||
if (utils.isArray(val)) {
|
||
key = key + '[]';
|
||
} else {
|
||
val = [val];
|
||
}
|
||
|
||
utils.forEach(val, function parseValue(v) {
|
||
if (utils.isDate(v)) {
|
||
v = v.toISOString();
|
||
} else if (utils.isObject(v)) {
|
||
v = JSON.stringify(v);
|
||
}
|
||
parts.push(encode(key) + '=' + encode(v));
|
||
});
|
||
});
|
||
|
||
serializedParams = parts.join('&');
|
||
}
|
||
|
||
if (serializedParams) {
|
||
var hashmarkIndex = url.indexOf('#');
|
||
if (hashmarkIndex !== -1) {
|
||
url = url.slice(0, hashmarkIndex);
|
||
}
|
||
|
||
url += (url.indexOf('?') === -1 ? '?' : '&') + serializedParams;
|
||
}
|
||
|
||
return url;
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 137:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var Cancel = __webpack_require__(826);
|
||
|
||
/**
|
||
* A `CancelToken` is an object that can be used to request cancellation of an operation.
|
||
*
|
||
* @class
|
||
* @param {Function} executor The executor function.
|
||
*/
|
||
function CancelToken(executor) {
|
||
if (typeof executor !== 'function') {
|
||
throw new TypeError('executor must be a function.');
|
||
}
|
||
|
||
var resolvePromise;
|
||
this.promise = new Promise(function promiseExecutor(resolve) {
|
||
resolvePromise = resolve;
|
||
});
|
||
|
||
var token = this;
|
||
executor(function cancel(message) {
|
||
if (token.reason) {
|
||
// Cancellation has already been requested
|
||
return;
|
||
}
|
||
|
||
token.reason = new Cancel(message);
|
||
resolvePromise(token.reason);
|
||
});
|
||
}
|
||
|
||
/**
|
||
* Throws a `Cancel` if cancellation has been requested.
|
||
*/
|
||
CancelToken.prototype.throwIfRequested = function throwIfRequested() {
|
||
if (this.reason) {
|
||
throw this.reason;
|
||
}
|
||
};
|
||
|
||
/**
|
||
* Returns an object that contains a new `CancelToken` and a function that, when called,
|
||
* cancels the `CancelToken`.
|
||
*/
|
||
CancelToken.source = function source() {
|
||
var cancel;
|
||
var token = new CancelToken(function executor(c) {
|
||
cancel = c;
|
||
});
|
||
return {
|
||
token: token,
|
||
cancel: cancel
|
||
};
|
||
};
|
||
|
||
module.exports = CancelToken;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 141:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var net = __webpack_require__(937);
|
||
var tls = __webpack_require__(16);
|
||
var http = __webpack_require__(605);
|
||
var https = __webpack_require__(211);
|
||
var events = __webpack_require__(614);
|
||
var assert = __webpack_require__(357);
|
||
var util = __webpack_require__(669);
|
||
|
||
|
||
exports.httpOverHttp = httpOverHttp;
|
||
exports.httpsOverHttp = httpsOverHttp;
|
||
exports.httpOverHttps = httpOverHttps;
|
||
exports.httpsOverHttps = httpsOverHttps;
|
||
|
||
|
||
function httpOverHttp(options) {
|
||
var agent = new TunnelingAgent(options);
|
||
agent.request = http.request;
|
||
return agent;
|
||
}
|
||
|
||
function httpsOverHttp(options) {
|
||
var agent = new TunnelingAgent(options);
|
||
agent.request = http.request;
|
||
agent.createSocket = createSecureSocket;
|
||
agent.defaultPort = 443;
|
||
return agent;
|
||
}
|
||
|
||
function httpOverHttps(options) {
|
||
var agent = new TunnelingAgent(options);
|
||
agent.request = https.request;
|
||
return agent;
|
||
}
|
||
|
||
function httpsOverHttps(options) {
|
||
var agent = new TunnelingAgent(options);
|
||
agent.request = https.request;
|
||
agent.createSocket = createSecureSocket;
|
||
agent.defaultPort = 443;
|
||
return agent;
|
||
}
|
||
|
||
|
||
function TunnelingAgent(options) {
|
||
var self = this;
|
||
self.options = options || {};
|
||
self.proxyOptions = self.options.proxy || {};
|
||
self.maxSockets = self.options.maxSockets || http.Agent.defaultMaxSockets;
|
||
self.requests = [];
|
||
self.sockets = [];
|
||
|
||
self.on('free', function onFree(socket, host, port, localAddress) {
|
||
var options = toOptions(host, port, localAddress);
|
||
for (var i = 0, len = self.requests.length; i < len; ++i) {
|
||
var pending = self.requests[i];
|
||
if (pending.host === options.host && pending.port === options.port) {
|
||
// Detect the request to connect same origin server,
|
||
// reuse the connection.
|
||
self.requests.splice(i, 1);
|
||
pending.request.onSocket(socket);
|
||
return;
|
||
}
|
||
}
|
||
socket.destroy();
|
||
self.removeSocket(socket);
|
||
});
|
||
}
|
||
util.inherits(TunnelingAgent, events.EventEmitter);
|
||
|
||
TunnelingAgent.prototype.addRequest = function addRequest(req, host, port, localAddress) {
|
||
var self = this;
|
||
var options = mergeOptions({request: req}, self.options, toOptions(host, port, localAddress));
|
||
|
||
if (self.sockets.length >= this.maxSockets) {
|
||
// We are over limit so we'll add it to the queue.
|
||
self.requests.push(options);
|
||
return;
|
||
}
|
||
|
||
// If we are under maxSockets create a new one.
|
||
self.createSocket(options, function(socket) {
|
||
socket.on('free', onFree);
|
||
socket.on('close', onCloseOrRemove);
|
||
socket.on('agentRemove', onCloseOrRemove);
|
||
req.onSocket(socket);
|
||
|
||
function onFree() {
|
||
self.emit('free', socket, options);
|
||
}
|
||
|
||
function onCloseOrRemove(err) {
|
||
self.removeSocket(socket);
|
||
socket.removeListener('free', onFree);
|
||
socket.removeListener('close', onCloseOrRemove);
|
||
socket.removeListener('agentRemove', onCloseOrRemove);
|
||
}
|
||
});
|
||
};
|
||
|
||
TunnelingAgent.prototype.createSocket = function createSocket(options, cb) {
|
||
var self = this;
|
||
var placeholder = {};
|
||
self.sockets.push(placeholder);
|
||
|
||
var connectOptions = mergeOptions({}, self.proxyOptions, {
|
||
method: 'CONNECT',
|
||
path: options.host + ':' + options.port,
|
||
agent: false,
|
||
headers: {
|
||
host: options.host + ':' + options.port
|
||
}
|
||
});
|
||
if (options.localAddress) {
|
||
connectOptions.localAddress = options.localAddress;
|
||
}
|
||
if (connectOptions.proxyAuth) {
|
||
connectOptions.headers = connectOptions.headers || {};
|
||
connectOptions.headers['Proxy-Authorization'] = 'Basic ' +
|
||
new Buffer(connectOptions.proxyAuth).toString('base64');
|
||
}
|
||
|
||
debug('making CONNECT request');
|
||
var connectReq = self.request(connectOptions);
|
||
connectReq.useChunkedEncodingByDefault = false; // for v0.6
|
||
connectReq.once('response', onResponse); // for v0.6
|
||
connectReq.once('upgrade', onUpgrade); // for v0.6
|
||
connectReq.once('connect', onConnect); // for v0.7 or later
|
||
connectReq.once('error', onError);
|
||
connectReq.end();
|
||
|
||
function onResponse(res) {
|
||
// Very hacky. This is necessary to avoid http-parser leaks.
|
||
res.upgrade = true;
|
||
}
|
||
|
||
function onUpgrade(res, socket, head) {
|
||
// Hacky.
|
||
process.nextTick(function() {
|
||
onConnect(res, socket, head);
|
||
});
|
||
}
|
||
|
||
function onConnect(res, socket, head) {
|
||
connectReq.removeAllListeners();
|
||
socket.removeAllListeners();
|
||
|
||
if (res.statusCode !== 200) {
|
||
debug('tunneling socket could not be established, statusCode=%d',
|
||
res.statusCode);
|
||
socket.destroy();
|
||
var error = new Error('tunneling socket could not be established, ' +
|
||
'statusCode=' + res.statusCode);
|
||
error.code = 'ECONNRESET';
|
||
options.request.emit('error', error);
|
||
self.removeSocket(placeholder);
|
||
return;
|
||
}
|
||
if (head.length > 0) {
|
||
debug('got illegal response body from proxy');
|
||
socket.destroy();
|
||
var error = new Error('got illegal response body from proxy');
|
||
error.code = 'ECONNRESET';
|
||
options.request.emit('error', error);
|
||
self.removeSocket(placeholder);
|
||
return;
|
||
}
|
||
debug('tunneling connection has established');
|
||
self.sockets[self.sockets.indexOf(placeholder)] = socket;
|
||
return cb(socket);
|
||
}
|
||
|
||
function onError(cause) {
|
||
connectReq.removeAllListeners();
|
||
|
||
debug('tunneling socket could not be established, cause=%s\n',
|
||
cause.message, cause.stack);
|
||
var error = new Error('tunneling socket could not be established, ' +
|
||
'cause=' + cause.message);
|
||
error.code = 'ECONNRESET';
|
||
options.request.emit('error', error);
|
||
self.removeSocket(placeholder);
|
||
}
|
||
};
|
||
|
||
TunnelingAgent.prototype.removeSocket = function removeSocket(socket) {
|
||
var pos = this.sockets.indexOf(socket)
|
||
if (pos === -1) {
|
||
return;
|
||
}
|
||
this.sockets.splice(pos, 1);
|
||
|
||
var pending = this.requests.shift();
|
||
if (pending) {
|
||
// If we have pending requests and a socket gets closed a new one
|
||
// needs to be created to take over in the pool for the one that closed.
|
||
this.createSocket(pending, function(socket) {
|
||
pending.request.onSocket(socket);
|
||
});
|
||
}
|
||
};
|
||
|
||
function createSecureSocket(options, cb) {
|
||
var self = this;
|
||
TunnelingAgent.prototype.createSocket.call(self, options, function(socket) {
|
||
var hostHeader = options.request.getHeader('host');
|
||
var tlsOptions = mergeOptions({}, self.options, {
|
||
socket: socket,
|
||
servername: hostHeader ? hostHeader.replace(/:.*$/, '') : options.host
|
||
});
|
||
|
||
// 0 is dummy port for v0.6
|
||
var secureSocket = tls.connect(0, tlsOptions);
|
||
self.sockets[self.sockets.indexOf(socket)] = secureSocket;
|
||
cb(secureSocket);
|
||
});
|
||
}
|
||
|
||
|
||
function toOptions(host, port, localAddress) {
|
||
if (typeof host === 'string') { // since v0.10
|
||
return {
|
||
host: host,
|
||
port: port,
|
||
localAddress: localAddress
|
||
};
|
||
}
|
||
return host; // for v0.11 or later
|
||
}
|
||
|
||
function mergeOptions(target) {
|
||
for (var i = 1, len = arguments.length; i < len; ++i) {
|
||
var overrides = arguments[i];
|
||
if (typeof overrides === 'object') {
|
||
var keys = Object.keys(overrides);
|
||
for (var j = 0, keyLen = keys.length; j < keyLen; ++j) {
|
||
var k = keys[j];
|
||
if (overrides[k] !== undefined) {
|
||
target[k] = overrides[k];
|
||
}
|
||
}
|
||
}
|
||
}
|
||
return target;
|
||
}
|
||
|
||
|
||
var debug;
|
||
if (process.env.NODE_DEBUG && /\btunnel\b/.test(process.env.NODE_DEBUG)) {
|
||
debug = function() {
|
||
var args = Array.prototype.slice.call(arguments);
|
||
if (typeof args[0] === 'string') {
|
||
args[0] = 'TUNNEL: ' + args[0];
|
||
} else {
|
||
args.unshift('TUNNEL:');
|
||
}
|
||
console.error.apply(console, args);
|
||
}
|
||
} else {
|
||
debug = function() {};
|
||
}
|
||
exports.debug = debug; // for test
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 150:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
/*!
|
||
* Tmp
|
||
*
|
||
* Copyright (c) 2011-2017 KARASZI Istvan <github@spam.raszi.hu>
|
||
*
|
||
* MIT Licensed
|
||
*/
|
||
|
||
/*
|
||
* Module dependencies.
|
||
*/
|
||
const fs = __webpack_require__(747);
|
||
const os = __webpack_require__(87);
|
||
const path = __webpack_require__(622);
|
||
const crypto = __webpack_require__(417);
|
||
const _c = fs.constants && os.constants ?
|
||
{ fs: fs.constants, os: os.constants } :
|
||
process.binding('constants');
|
||
const rimraf = __webpack_require__(569);
|
||
|
||
/*
|
||
* The working inner variables.
|
||
*/
|
||
const
|
||
// the random characters to choose from
|
||
RANDOM_CHARS = '0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz',
|
||
|
||
TEMPLATE_PATTERN = /XXXXXX/,
|
||
|
||
DEFAULT_TRIES = 3,
|
||
|
||
CREATE_FLAGS = (_c.O_CREAT || _c.fs.O_CREAT) | (_c.O_EXCL || _c.fs.O_EXCL) | (_c.O_RDWR || _c.fs.O_RDWR),
|
||
|
||
EBADF = _c.EBADF || _c.os.errno.EBADF,
|
||
ENOENT = _c.ENOENT || _c.os.errno.ENOENT,
|
||
|
||
DIR_MODE = 448 /* 0o700 */,
|
||
FILE_MODE = 384 /* 0o600 */,
|
||
|
||
EXIT = 'exit',
|
||
|
||
SIGINT = 'SIGINT',
|
||
|
||
// this will hold the objects need to be removed on exit
|
||
_removeObjects = [];
|
||
|
||
var
|
||
_gracefulCleanup = false;
|
||
|
||
/**
|
||
* Random name generator based on crypto.
|
||
* Adapted from http://blog.tompawlak.org/how-to-generate-random-values-nodejs-javascript
|
||
*
|
||
* @param {number} howMany
|
||
* @returns {string} the generated random name
|
||
* @private
|
||
*/
|
||
function _randomChars(howMany) {
|
||
var
|
||
value = [],
|
||
rnd = null;
|
||
|
||
// make sure that we do not fail because we ran out of entropy
|
||
try {
|
||
rnd = crypto.randomBytes(howMany);
|
||
} catch (e) {
|
||
rnd = crypto.pseudoRandomBytes(howMany);
|
||
}
|
||
|
||
for (var i = 0; i < howMany; i++) {
|
||
value.push(RANDOM_CHARS[rnd[i] % RANDOM_CHARS.length]);
|
||
}
|
||
|
||
return value.join('');
|
||
}
|
||
|
||
/**
|
||
* Checks whether the `obj` parameter is defined or not.
|
||
*
|
||
* @param {Object} obj
|
||
* @returns {boolean} true if the object is undefined
|
||
* @private
|
||
*/
|
||
function _isUndefined(obj) {
|
||
return typeof obj === 'undefined';
|
||
}
|
||
|
||
/**
|
||
* Parses the function arguments.
|
||
*
|
||
* This function helps to have optional arguments.
|
||
*
|
||
* @param {(Options|Function)} options
|
||
* @param {Function} callback
|
||
* @returns {Array} parsed arguments
|
||
* @private
|
||
*/
|
||
function _parseArguments(options, callback) {
|
||
/* istanbul ignore else */
|
||
if (typeof options === 'function') {
|
||
return [{}, options];
|
||
}
|
||
|
||
/* istanbul ignore else */
|
||
if (_isUndefined(options)) {
|
||
return [{}, callback];
|
||
}
|
||
|
||
return [options, callback];
|
||
}
|
||
|
||
/**
|
||
* Generates a new temporary name.
|
||
*
|
||
* @param {Object} opts
|
||
* @returns {string} the new random name according to opts
|
||
* @private
|
||
*/
|
||
function _generateTmpName(opts) {
|
||
|
||
const tmpDir = _getTmpDir();
|
||
|
||
// fail early on missing tmp dir
|
||
if (isBlank(opts.dir) && isBlank(tmpDir)) {
|
||
throw new Error('No tmp dir specified');
|
||
}
|
||
|
||
/* istanbul ignore else */
|
||
if (!isBlank(opts.name)) {
|
||
return path.join(opts.dir || tmpDir, opts.name);
|
||
}
|
||
|
||
// mkstemps like template
|
||
// opts.template has already been guarded in tmpName() below
|
||
/* istanbul ignore else */
|
||
if (opts.template) {
|
||
var template = opts.template;
|
||
// make sure that we prepend the tmp path if none was given
|
||
/* istanbul ignore else */
|
||
if (path.basename(template) === template)
|
||
template = path.join(opts.dir || tmpDir, template);
|
||
return template.replace(TEMPLATE_PATTERN, _randomChars(6));
|
||
}
|
||
|
||
// prefix and postfix
|
||
const name = [
|
||
(isBlank(opts.prefix) ? 'tmp-' : opts.prefix),
|
||
process.pid,
|
||
_randomChars(12),
|
||
(opts.postfix ? opts.postfix : '')
|
||
].join('');
|
||
|
||
return path.join(opts.dir || tmpDir, name);
|
||
}
|
||
|
||
/**
|
||
* Gets a temporary file name.
|
||
*
|
||
* @param {(Options|tmpNameCallback)} options options or callback
|
||
* @param {?tmpNameCallback} callback the callback function
|
||
*/
|
||
function tmpName(options, callback) {
|
||
var
|
||
args = _parseArguments(options, callback),
|
||
opts = args[0],
|
||
cb = args[1],
|
||
tries = !isBlank(opts.name) ? 1 : opts.tries || DEFAULT_TRIES;
|
||
|
||
/* istanbul ignore else */
|
||
if (isNaN(tries) || tries < 0)
|
||
return cb(new Error('Invalid tries'));
|
||
|
||
/* istanbul ignore else */
|
||
if (opts.template && !opts.template.match(TEMPLATE_PATTERN))
|
||
return cb(new Error('Invalid template provided'));
|
||
|
||
(function _getUniqueName() {
|
||
try {
|
||
const name = _generateTmpName(opts);
|
||
|
||
// check whether the path exists then retry if needed
|
||
fs.stat(name, function (err) {
|
||
/* istanbul ignore else */
|
||
if (!err) {
|
||
/* istanbul ignore else */
|
||
if (tries-- > 0) return _getUniqueName();
|
||
|
||
return cb(new Error('Could not get a unique tmp filename, max tries reached ' + name));
|
||
}
|
||
|
||
cb(null, name);
|
||
});
|
||
} catch (err) {
|
||
cb(err);
|
||
}
|
||
}());
|
||
}
|
||
|
||
/**
|
||
* Synchronous version of tmpName.
|
||
*
|
||
* @param {Object} options
|
||
* @returns {string} the generated random name
|
||
* @throws {Error} if the options are invalid or could not generate a filename
|
||
*/
|
||
function tmpNameSync(options) {
|
||
var
|
||
args = _parseArguments(options),
|
||
opts = args[0],
|
||
tries = !isBlank(opts.name) ? 1 : opts.tries || DEFAULT_TRIES;
|
||
|
||
/* istanbul ignore else */
|
||
if (isNaN(tries) || tries < 0)
|
||
throw new Error('Invalid tries');
|
||
|
||
/* istanbul ignore else */
|
||
if (opts.template && !opts.template.match(TEMPLATE_PATTERN))
|
||
throw new Error('Invalid template provided');
|
||
|
||
do {
|
||
const name = _generateTmpName(opts);
|
||
try {
|
||
fs.statSync(name);
|
||
} catch (e) {
|
||
return name;
|
||
}
|
||
} while (tries-- > 0);
|
||
|
||
throw new Error('Could not get a unique tmp filename, max tries reached');
|
||
}
|
||
|
||
/**
|
||
* Creates and opens a temporary file.
|
||
*
|
||
* @param {(Options|fileCallback)} options the config options or the callback function
|
||
* @param {?fileCallback} callback
|
||
*/
|
||
function file(options, callback) {
|
||
var
|
||
args = _parseArguments(options, callback),
|
||
opts = args[0],
|
||
cb = args[1];
|
||
|
||
// gets a temporary filename
|
||
tmpName(opts, function _tmpNameCreated(err, name) {
|
||
/* istanbul ignore else */
|
||
if (err) return cb(err);
|
||
|
||
// create and open the file
|
||
fs.open(name, CREATE_FLAGS, opts.mode || FILE_MODE, function _fileCreated(err, fd) {
|
||
/* istanbul ignore else */
|
||
if (err) return cb(err);
|
||
|
||
if (opts.discardDescriptor) {
|
||
return fs.close(fd, function _discardCallback(err) {
|
||
/* istanbul ignore else */
|
||
if (err) {
|
||
// Low probability, and the file exists, so this could be
|
||
// ignored. If it isn't we certainly need to unlink the
|
||
// file, and if that fails too its error is more
|
||
// important.
|
||
try {
|
||
fs.unlinkSync(name);
|
||
} catch (e) {
|
||
if (!isENOENT(e)) {
|
||
err = e;
|
||
}
|
||
}
|
||
return cb(err);
|
||
}
|
||
cb(null, name, undefined, _prepareTmpFileRemoveCallback(name, -1, opts));
|
||
});
|
||
}
|
||
/* istanbul ignore else */
|
||
if (opts.detachDescriptor) {
|
||
return cb(null, name, fd, _prepareTmpFileRemoveCallback(name, -1, opts));
|
||
}
|
||
cb(null, name, fd, _prepareTmpFileRemoveCallback(name, fd, opts));
|
||
});
|
||
});
|
||
}
|
||
|
||
/**
|
||
* Synchronous version of file.
|
||
*
|
||
* @param {Options} options
|
||
* @returns {FileSyncObject} object consists of name, fd and removeCallback
|
||
* @throws {Error} if cannot create a file
|
||
*/
|
||
function fileSync(options) {
|
||
var
|
||
args = _parseArguments(options),
|
||
opts = args[0];
|
||
|
||
const discardOrDetachDescriptor = opts.discardDescriptor || opts.detachDescriptor;
|
||
const name = tmpNameSync(opts);
|
||
var fd = fs.openSync(name, CREATE_FLAGS, opts.mode || FILE_MODE);
|
||
/* istanbul ignore else */
|
||
if (opts.discardDescriptor) {
|
||
fs.closeSync(fd);
|
||
fd = undefined;
|
||
}
|
||
|
||
return {
|
||
name: name,
|
||
fd: fd,
|
||
removeCallback: _prepareTmpFileRemoveCallback(name, discardOrDetachDescriptor ? -1 : fd, opts)
|
||
};
|
||
}
|
||
|
||
/**
|
||
* Creates a temporary directory.
|
||
*
|
||
* @param {(Options|dirCallback)} options the options or the callback function
|
||
* @param {?dirCallback} callback
|
||
*/
|
||
function dir(options, callback) {
|
||
var
|
||
args = _parseArguments(options, callback),
|
||
opts = args[0],
|
||
cb = args[1];
|
||
|
||
// gets a temporary filename
|
||
tmpName(opts, function _tmpNameCreated(err, name) {
|
||
/* istanbul ignore else */
|
||
if (err) return cb(err);
|
||
|
||
// create the directory
|
||
fs.mkdir(name, opts.mode || DIR_MODE, function _dirCreated(err) {
|
||
/* istanbul ignore else */
|
||
if (err) return cb(err);
|
||
|
||
cb(null, name, _prepareTmpDirRemoveCallback(name, opts));
|
||
});
|
||
});
|
||
}
|
||
|
||
/**
|
||
* Synchronous version of dir.
|
||
*
|
||
* @param {Options} options
|
||
* @returns {DirSyncObject} object consists of name and removeCallback
|
||
* @throws {Error} if it cannot create a directory
|
||
*/
|
||
function dirSync(options) {
|
||
var
|
||
args = _parseArguments(options),
|
||
opts = args[0];
|
||
|
||
const name = tmpNameSync(opts);
|
||
fs.mkdirSync(name, opts.mode || DIR_MODE);
|
||
|
||
return {
|
||
name: name,
|
||
removeCallback: _prepareTmpDirRemoveCallback(name, opts)
|
||
};
|
||
}
|
||
|
||
/**
|
||
* Removes files asynchronously.
|
||
*
|
||
* @param {Object} fdPath
|
||
* @param {Function} next
|
||
* @private
|
||
*/
|
||
function _removeFileAsync(fdPath, next) {
|
||
const _handler = function (err) {
|
||
if (err && !isENOENT(err)) {
|
||
// reraise any unanticipated error
|
||
return next(err);
|
||
}
|
||
next();
|
||
}
|
||
|
||
if (0 <= fdPath[0])
|
||
fs.close(fdPath[0], function (err) {
|
||
fs.unlink(fdPath[1], _handler);
|
||
});
|
||
else fs.unlink(fdPath[1], _handler);
|
||
}
|
||
|
||
/**
|
||
* Removes files synchronously.
|
||
*
|
||
* @param {Object} fdPath
|
||
* @private
|
||
*/
|
||
function _removeFileSync(fdPath) {
|
||
try {
|
||
if (0 <= fdPath[0]) fs.closeSync(fdPath[0]);
|
||
} catch (e) {
|
||
// reraise any unanticipated error
|
||
if (!isEBADF(e) && !isENOENT(e)) throw e;
|
||
} finally {
|
||
try {
|
||
fs.unlinkSync(fdPath[1]);
|
||
}
|
||
catch (e) {
|
||
// reraise any unanticipated error
|
||
if (!isENOENT(e)) throw e;
|
||
}
|
||
}
|
||
}
|
||
|
||
/**
|
||
* Prepares the callback for removal of the temporary file.
|
||
*
|
||
* @param {string} name the path of the file
|
||
* @param {number} fd file descriptor
|
||
* @param {Object} opts
|
||
* @returns {fileCallback}
|
||
* @private
|
||
*/
|
||
function _prepareTmpFileRemoveCallback(name, fd, opts) {
|
||
const removeCallbackSync = _prepareRemoveCallback(_removeFileSync, [fd, name]);
|
||
const removeCallback = _prepareRemoveCallback(_removeFileAsync, [fd, name], removeCallbackSync);
|
||
|
||
if (!opts.keep) _removeObjects.unshift(removeCallbackSync);
|
||
|
||
return removeCallback;
|
||
}
|
||
|
||
/**
|
||
* Simple wrapper for rimraf.
|
||
*
|
||
* @param {string} dirPath
|
||
* @param {Function} next
|
||
* @private
|
||
*/
|
||
function _rimrafRemoveDirWrapper(dirPath, next) {
|
||
rimraf(dirPath, next);
|
||
}
|
||
|
||
/**
|
||
* Simple wrapper for rimraf.sync.
|
||
*
|
||
* @param {string} dirPath
|
||
* @private
|
||
*/
|
||
function _rimrafRemoveDirSyncWrapper(dirPath, next) {
|
||
try {
|
||
return next(null, rimraf.sync(dirPath));
|
||
} catch (err) {
|
||
return next(err);
|
||
}
|
||
}
|
||
|
||
/**
|
||
* Prepares the callback for removal of the temporary directory.
|
||
*
|
||
* @param {string} name
|
||
* @param {Object} opts
|
||
* @returns {Function} the callback
|
||
* @private
|
||
*/
|
||
function _prepareTmpDirRemoveCallback(name, opts) {
|
||
const removeFunction = opts.unsafeCleanup ? _rimrafRemoveDirWrapper : fs.rmdir.bind(fs);
|
||
const removeFunctionSync = opts.unsafeCleanup ? _rimrafRemoveDirSyncWrapper : fs.rmdirSync.bind(fs);
|
||
const removeCallbackSync = _prepareRemoveCallback(removeFunctionSync, name);
|
||
const removeCallback = _prepareRemoveCallback(removeFunction, name, removeCallbackSync);
|
||
if (!opts.keep) _removeObjects.unshift(removeCallbackSync);
|
||
|
||
return removeCallback;
|
||
}
|
||
|
||
/**
|
||
* Creates a guarded function wrapping the removeFunction call.
|
||
*
|
||
* @param {Function} removeFunction
|
||
* @param {Object} arg
|
||
* @returns {Function}
|
||
* @private
|
||
*/
|
||
function _prepareRemoveCallback(removeFunction, arg, cleanupCallbackSync) {
|
||
var called = false;
|
||
|
||
return function _cleanupCallback(next) {
|
||
next = next || function () {};
|
||
if (!called) {
|
||
const toRemove = cleanupCallbackSync || _cleanupCallback;
|
||
const index = _removeObjects.indexOf(toRemove);
|
||
/* istanbul ignore else */
|
||
if (index >= 0) _removeObjects.splice(index, 1);
|
||
|
||
called = true;
|
||
// sync?
|
||
if (removeFunction.length === 1) {
|
||
try {
|
||
removeFunction(arg);
|
||
return next(null);
|
||
}
|
||
catch (err) {
|
||
// if no next is provided and since we are
|
||
// in silent cleanup mode on process exit,
|
||
// we will ignore the error
|
||
return next(err);
|
||
}
|
||
} else return removeFunction(arg, next);
|
||
} else return next(new Error('cleanup callback has already been called'));
|
||
};
|
||
}
|
||
|
||
/**
|
||
* The garbage collector.
|
||
*
|
||
* @private
|
||
*/
|
||
function _garbageCollector() {
|
||
/* istanbul ignore else */
|
||
if (!_gracefulCleanup) return;
|
||
|
||
// the function being called removes itself from _removeObjects,
|
||
// loop until _removeObjects is empty
|
||
while (_removeObjects.length) {
|
||
try {
|
||
_removeObjects[0]();
|
||
} catch (e) {
|
||
// already removed?
|
||
}
|
||
}
|
||
}
|
||
|
||
/**
|
||
* Helper for testing against EBADF to compensate changes made to Node 7.x under Windows.
|
||
*/
|
||
function isEBADF(error) {
|
||
return isExpectedError(error, -EBADF, 'EBADF');
|
||
}
|
||
|
||
/**
|
||
* Helper for testing against ENOENT to compensate changes made to Node 7.x under Windows.
|
||
*/
|
||
function isENOENT(error) {
|
||
return isExpectedError(error, -ENOENT, 'ENOENT');
|
||
}
|
||
|
||
/**
|
||
* Helper to determine whether the expected error code matches the actual code and errno,
|
||
* which will differ between the supported node versions.
|
||
*
|
||
* - Node >= 7.0:
|
||
* error.code {string}
|
||
* error.errno {string|number} any numerical value will be negated
|
||
*
|
||
* - Node >= 6.0 < 7.0:
|
||
* error.code {string}
|
||
* error.errno {number} negated
|
||
*
|
||
* - Node >= 4.0 < 6.0: introduces SystemError
|
||
* error.code {string}
|
||
* error.errno {number} negated
|
||
*
|
||
* - Node >= 0.10 < 4.0:
|
||
* error.code {number} negated
|
||
* error.errno n/a
|
||
*/
|
||
function isExpectedError(error, code, errno) {
|
||
return error.code === code || error.code === errno;
|
||
}
|
||
|
||
/**
|
||
* Helper which determines whether a string s is blank, that is undefined, or empty or null.
|
||
*
|
||
* @private
|
||
* @param {string} s
|
||
* @returns {Boolean} true whether the string s is blank, false otherwise
|
||
*/
|
||
function isBlank(s) {
|
||
return s === null || s === undefined || !s.trim();
|
||
}
|
||
|
||
/**
|
||
* Sets the graceful cleanup.
|
||
*/
|
||
function setGracefulCleanup() {
|
||
_gracefulCleanup = true;
|
||
}
|
||
|
||
/**
|
||
* Returns the currently configured tmp dir from os.tmpdir().
|
||
*
|
||
* @private
|
||
* @returns {string} the currently configured tmp dir
|
||
*/
|
||
function _getTmpDir() {
|
||
return os.tmpdir();
|
||
}
|
||
|
||
/**
|
||
* If there are multiple different versions of tmp in place, make sure that
|
||
* we recognize the old listeners.
|
||
*
|
||
* @param {Function} listener
|
||
* @private
|
||
* @returns {Boolean} true whether listener is a legacy listener
|
||
*/
|
||
function _is_legacy_listener(listener) {
|
||
return (listener.name === '_exit' || listener.name === '_uncaughtExceptionThrown')
|
||
&& listener.toString().indexOf('_garbageCollector();') > -1;
|
||
}
|
||
|
||
/**
|
||
* Safely install SIGINT listener.
|
||
*
|
||
* NOTE: this will only work on OSX and Linux.
|
||
*
|
||
* @private
|
||
*/
|
||
function _safely_install_sigint_listener() {
|
||
|
||
const listeners = process.listeners(SIGINT);
|
||
const existingListeners = [];
|
||
for (let i = 0, length = listeners.length; i < length; i++) {
|
||
const lstnr = listeners[i];
|
||
/* istanbul ignore else */
|
||
if (lstnr.name === '_tmp$sigint_listener') {
|
||
existingListeners.push(lstnr);
|
||
process.removeListener(SIGINT, lstnr);
|
||
}
|
||
}
|
||
process.on(SIGINT, function _tmp$sigint_listener(doExit) {
|
||
for (let i = 0, length = existingListeners.length; i < length; i++) {
|
||
// let the existing listener do the garbage collection (e.g. jest sandbox)
|
||
try {
|
||
existingListeners[i](false);
|
||
} catch (err) {
|
||
// ignore
|
||
}
|
||
}
|
||
try {
|
||
// force the garbage collector even it is called again in the exit listener
|
||
_garbageCollector();
|
||
} finally {
|
||
if (!!doExit) {
|
||
process.exit(0);
|
||
}
|
||
}
|
||
});
|
||
}
|
||
|
||
/**
|
||
* Safely install process exit listener.
|
||
*
|
||
* @private
|
||
*/
|
||
function _safely_install_exit_listener() {
|
||
const listeners = process.listeners(EXIT);
|
||
|
||
// collect any existing listeners
|
||
const existingListeners = [];
|
||
for (let i = 0, length = listeners.length; i < length; i++) {
|
||
const lstnr = listeners[i];
|
||
/* istanbul ignore else */
|
||
// TODO: remove support for legacy listeners once release 1.0.0 is out
|
||
if (lstnr.name === '_tmp$safe_listener' || _is_legacy_listener(lstnr)) {
|
||
// we must forget about the uncaughtException listener, hopefully it is ours
|
||
if (lstnr.name !== '_uncaughtExceptionThrown') {
|
||
existingListeners.push(lstnr);
|
||
}
|
||
process.removeListener(EXIT, lstnr);
|
||
}
|
||
}
|
||
// TODO: what was the data parameter good for?
|
||
process.addListener(EXIT, function _tmp$safe_listener(data) {
|
||
for (let i = 0, length = existingListeners.length; i < length; i++) {
|
||
// let the existing listener do the garbage collection (e.g. jest sandbox)
|
||
try {
|
||
existingListeners[i](data);
|
||
} catch (err) {
|
||
// ignore
|
||
}
|
||
}
|
||
_garbageCollector();
|
||
});
|
||
}
|
||
|
||
_safely_install_exit_listener();
|
||
_safely_install_sigint_listener();
|
||
|
||
/**
|
||
* Configuration options.
|
||
*
|
||
* @typedef {Object} Options
|
||
* @property {?number} tries the number of tries before give up the name generation
|
||
* @property {?string} template the "mkstemp" like filename template
|
||
* @property {?string} name fix name
|
||
* @property {?string} dir the tmp directory to use
|
||
* @property {?string} prefix prefix for the generated name
|
||
* @property {?string} postfix postfix for the generated name
|
||
* @property {?boolean} unsafeCleanup recursively removes the created temporary directory, even when it's not empty
|
||
*/
|
||
|
||
/**
|
||
* @typedef {Object} FileSyncObject
|
||
* @property {string} name the name of the file
|
||
* @property {string} fd the file descriptor
|
||
* @property {fileCallback} removeCallback the callback function to remove the file
|
||
*/
|
||
|
||
/**
|
||
* @typedef {Object} DirSyncObject
|
||
* @property {string} name the name of the directory
|
||
* @property {fileCallback} removeCallback the callback function to remove the directory
|
||
*/
|
||
|
||
/**
|
||
* @callback tmpNameCallback
|
||
* @param {?Error} err the error object if anything goes wrong
|
||
* @param {string} name the temporary file name
|
||
*/
|
||
|
||
/**
|
||
* @callback fileCallback
|
||
* @param {?Error} err the error object if anything goes wrong
|
||
* @param {string} name the temporary file name
|
||
* @param {number} fd the file descriptor
|
||
* @param {cleanupCallback} fn the cleanup callback function
|
||
*/
|
||
|
||
/**
|
||
* @callback dirCallback
|
||
* @param {?Error} err the error object if anything goes wrong
|
||
* @param {string} name the temporary file name
|
||
* @param {cleanupCallback} fn the cleanup callback function
|
||
*/
|
||
|
||
/**
|
||
* Removes the temporary created file or directory.
|
||
*
|
||
* @callback cleanupCallback
|
||
* @param {simpleCallback} [next] function to call after entry was removed
|
||
*/
|
||
|
||
/**
|
||
* Callback function for function composition.
|
||
* @see {@link https://github.com/raszi/node-tmp/issues/57|raszi/node-tmp#57}
|
||
*
|
||
* @callback simpleCallback
|
||
*/
|
||
|
||
// exporting all the needed methods
|
||
|
||
// evaluate os.tmpdir() lazily, mainly for simplifying testing but it also will
|
||
// allow users to reconfigure the temporary directory
|
||
Object.defineProperty(module.exports, 'tmpdir', {
|
||
enumerable: true,
|
||
configurable: false,
|
||
get: function () {
|
||
return _getTmpDir();
|
||
}
|
||
});
|
||
|
||
module.exports.dir = dir;
|
||
module.exports.dirSync = dirSync;
|
||
|
||
module.exports.file = file;
|
||
module.exports.fileSync = fileSync;
|
||
|
||
module.exports.tmpName = tmpName;
|
||
module.exports.tmpNameSync = tmpNameSync;
|
||
|
||
module.exports.setGracefulCleanup = setGracefulCleanup;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 176:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const core_1 = __webpack_require__(470);
|
||
/**
|
||
* Status Reporter that displays information about the progress/status of an artifact that is being uploaded or downloaded
|
||
*
|
||
* Variable display time that can be adjusted using the displayFrequencyInMilliseconds variable
|
||
* The total status of the upload/download gets displayed according to this value
|
||
* If there is a large file that is being uploaded, extra information about the individual status can also be displayed using the updateLargeFileStatus function
|
||
*/
|
||
class StatusReporter {
|
||
constructor(displayFrequencyInMilliseconds) {
|
||
this.totalNumberOfFilesToProcess = 0;
|
||
this.processedCount = 0;
|
||
this.largeFiles = new Map();
|
||
this.totalFileStatus = undefined;
|
||
this.largeFileStatus = undefined;
|
||
this.displayFrequencyInMilliseconds = displayFrequencyInMilliseconds;
|
||
}
|
||
setTotalNumberOfFilesToProcess(fileTotal) {
|
||
this.totalNumberOfFilesToProcess = fileTotal;
|
||
}
|
||
start() {
|
||
// displays information about the total upload/download status
|
||
this.totalFileStatus = setInterval(() => {
|
||
// display 1 decimal place without any rounding
|
||
const percentage = this.formatPercentage(this.processedCount, this.totalNumberOfFilesToProcess);
|
||
core_1.info(`Total file count: ${this.totalNumberOfFilesToProcess} ---- Processed file #${this.processedCount} (${percentage.slice(0, percentage.indexOf('.') + 2)}%)`);
|
||
}, this.displayFrequencyInMilliseconds);
|
||
// displays extra information about any large files that take a significant amount of time to upload or download every 1 second
|
||
this.largeFileStatus = setInterval(() => {
|
||
for (const value of Array.from(this.largeFiles.values())) {
|
||
core_1.info(value);
|
||
}
|
||
// delete all entries in the map after displaying the information so it will not be displayed again unless explicitly added
|
||
this.largeFiles.clear();
|
||
}, 1000);
|
||
}
|
||
// if there is a large file that is being uploaded in chunks, this is used to display extra information about the status of the upload
|
||
updateLargeFileStatus(fileName, numerator, denominator) {
|
||
// display 1 decimal place without any rounding
|
||
const percentage = this.formatPercentage(numerator, denominator);
|
||
const displayInformation = `Uploading ${fileName} (${percentage.slice(0, percentage.indexOf('.') + 2)}%)`;
|
||
// any previously added display information should be overwritten for the specific large file because a map is being used
|
||
this.largeFiles.set(fileName, displayInformation);
|
||
}
|
||
stop() {
|
||
if (this.totalFileStatus) {
|
||
clearInterval(this.totalFileStatus);
|
||
}
|
||
if (this.largeFileStatus) {
|
||
clearInterval(this.largeFileStatus);
|
||
}
|
||
}
|
||
incrementProcessedCount() {
|
||
this.processedCount++;
|
||
}
|
||
formatPercentage(numerator, denominator) {
|
||
// toFixed() rounds, so use extra precision to display accurate information even though 4 decimal places are not displayed
|
||
return ((numerator / denominator) * 100).toFixed(4).toString();
|
||
}
|
||
}
|
||
exports.StatusReporter = StatusReporter;
|
||
//# sourceMappingURL=status-reporter.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 211:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("https");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 214:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const artifact_client_1 = __webpack_require__(359);
|
||
/**
|
||
* Constructs an ArtifactClient
|
||
*/
|
||
function create() {
|
||
return artifact_client_1.DefaultArtifactClient.create();
|
||
}
|
||
exports.create = create;
|
||
//# sourceMappingURL=artifact-client.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 219:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var utils = __webpack_require__(35);
|
||
var settle = __webpack_require__(564);
|
||
var cookies = __webpack_require__(864);
|
||
var buildURL = __webpack_require__(133);
|
||
var buildFullPath = __webpack_require__(960);
|
||
var parseHeaders = __webpack_require__(631);
|
||
var isURLSameOrigin = __webpack_require__(688);
|
||
var createError = __webpack_require__(26);
|
||
|
||
module.exports = function xhrAdapter(config) {
|
||
return new Promise(function dispatchXhrRequest(resolve, reject) {
|
||
var requestData = config.data;
|
||
var requestHeaders = config.headers;
|
||
var responseType = config.responseType;
|
||
|
||
if (utils.isFormData(requestData)) {
|
||
delete requestHeaders['Content-Type']; // Let the browser set it
|
||
}
|
||
|
||
var request = new XMLHttpRequest();
|
||
|
||
// HTTP basic authentication
|
||
if (config.auth) {
|
||
var username = config.auth.username || '';
|
||
var password = config.auth.password ? unescape(encodeURIComponent(config.auth.password)) : '';
|
||
requestHeaders.Authorization = 'Basic ' + btoa(username + ':' + password);
|
||
}
|
||
|
||
var fullPath = buildFullPath(config.baseURL, config.url);
|
||
request.open(config.method.toUpperCase(), buildURL(fullPath, config.params, config.paramsSerializer), true);
|
||
|
||
// Set the request timeout in MS
|
||
request.timeout = config.timeout;
|
||
|
||
function onloadend() {
|
||
if (!request) {
|
||
return;
|
||
}
|
||
// Prepare the response
|
||
var responseHeaders = 'getAllResponseHeaders' in request ? parseHeaders(request.getAllResponseHeaders()) : null;
|
||
var responseData = !responseType || responseType === 'text' || responseType === 'json' ?
|
||
request.responseText : request.response;
|
||
var response = {
|
||
data: responseData,
|
||
status: request.status,
|
||
statusText: request.statusText,
|
||
headers: responseHeaders,
|
||
config: config,
|
||
request: request
|
||
};
|
||
|
||
settle(resolve, reject, response);
|
||
|
||
// Clean up request
|
||
request = null;
|
||
}
|
||
|
||
if ('onloadend' in request) {
|
||
// Use onloadend if available
|
||
request.onloadend = onloadend;
|
||
} else {
|
||
// Listen for ready state to emulate onloadend
|
||
request.onreadystatechange = function handleLoad() {
|
||
if (!request || request.readyState !== 4) {
|
||
return;
|
||
}
|
||
|
||
// The request errored out and we didn't get a response, this will be
|
||
// handled by onerror instead
|
||
// With one exception: request that using file: protocol, most browsers
|
||
// will return status as 0 even though it's a successful request
|
||
if (request.status === 0 && !(request.responseURL && request.responseURL.indexOf('file:') === 0)) {
|
||
return;
|
||
}
|
||
// readystate handler is calling before onerror or ontimeout handlers,
|
||
// so we should call onloadend on the next 'tick'
|
||
setTimeout(onloadend);
|
||
};
|
||
}
|
||
|
||
// Handle browser request cancellation (as opposed to a manual cancellation)
|
||
request.onabort = function handleAbort() {
|
||
if (!request) {
|
||
return;
|
||
}
|
||
|
||
reject(createError('Request aborted', config, 'ECONNABORTED', request));
|
||
|
||
// Clean up request
|
||
request = null;
|
||
};
|
||
|
||
// Handle low level network errors
|
||
request.onerror = function handleError() {
|
||
// Real errors are hidden from us by the browser
|
||
// onerror should only fire if it's a network error
|
||
reject(createError('Network Error', config, null, request));
|
||
|
||
// Clean up request
|
||
request = null;
|
||
};
|
||
|
||
// Handle timeout
|
||
request.ontimeout = function handleTimeout() {
|
||
var timeoutErrorMessage = 'timeout of ' + config.timeout + 'ms exceeded';
|
||
if (config.timeoutErrorMessage) {
|
||
timeoutErrorMessage = config.timeoutErrorMessage;
|
||
}
|
||
reject(createError(
|
||
timeoutErrorMessage,
|
||
config,
|
||
config.transitional && config.transitional.clarifyTimeoutError ? 'ETIMEDOUT' : 'ECONNABORTED',
|
||
request));
|
||
|
||
// Clean up request
|
||
request = null;
|
||
};
|
||
|
||
// Add xsrf header
|
||
// This is only done if running in a standard browser environment.
|
||
// Specifically not if we're in a web worker, or react-native.
|
||
if (utils.isStandardBrowserEnv()) {
|
||
// Add xsrf header
|
||
var xsrfValue = (config.withCredentials || isURLSameOrigin(fullPath)) && config.xsrfCookieName ?
|
||
cookies.read(config.xsrfCookieName) :
|
||
undefined;
|
||
|
||
if (xsrfValue) {
|
||
requestHeaders[config.xsrfHeaderName] = xsrfValue;
|
||
}
|
||
}
|
||
|
||
// Add headers to the request
|
||
if ('setRequestHeader' in request) {
|
||
utils.forEach(requestHeaders, function setRequestHeader(val, key) {
|
||
if (typeof requestData === 'undefined' && key.toLowerCase() === 'content-type') {
|
||
// Remove Content-Type if data is undefined
|
||
delete requestHeaders[key];
|
||
} else {
|
||
// Otherwise add header to the request
|
||
request.setRequestHeader(key, val);
|
||
}
|
||
});
|
||
}
|
||
|
||
// Add withCredentials to request if needed
|
||
if (!utils.isUndefined(config.withCredentials)) {
|
||
request.withCredentials = !!config.withCredentials;
|
||
}
|
||
|
||
// Add responseType to request if needed
|
||
if (responseType && responseType !== 'json') {
|
||
request.responseType = config.responseType;
|
||
}
|
||
|
||
// Handle progress if needed
|
||
if (typeof config.onDownloadProgress === 'function') {
|
||
request.addEventListener('progress', config.onDownloadProgress);
|
||
}
|
||
|
||
// Not all browsers support upload events
|
||
if (typeof config.onUploadProgress === 'function' && request.upload) {
|
||
request.upload.addEventListener('progress', config.onUploadProgress);
|
||
}
|
||
|
||
if (config.cancelToken) {
|
||
// Handle cancellation
|
||
config.cancelToken.promise.then(function onCanceled(cancel) {
|
||
if (!request) {
|
||
return;
|
||
}
|
||
|
||
request.abort();
|
||
reject(cancel);
|
||
// Clean up request
|
||
request = null;
|
||
});
|
||
}
|
||
|
||
if (!requestData) {
|
||
requestData = null;
|
||
}
|
||
|
||
// Send the request
|
||
request.send(requestData);
|
||
});
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 226:
|
||
/***/ (function(__unusedmodule, exports) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
class BasicCredentialHandler {
|
||
constructor(username, password) {
|
||
this.username = username;
|
||
this.password = password;
|
||
}
|
||
prepareRequest(options) {
|
||
options.headers['Authorization'] =
|
||
'Basic ' +
|
||
Buffer.from(this.username + ':' + this.password).toString('base64');
|
||
}
|
||
// This handler cannot handle 401
|
||
canHandleAuthentication(response) {
|
||
return false;
|
||
}
|
||
handleAuthentication(httpClient, requestInfo, objs) {
|
||
return null;
|
||
}
|
||
}
|
||
exports.BasicCredentialHandler = BasicCredentialHandler;
|
||
class BearerCredentialHandler {
|
||
constructor(token) {
|
||
this.token = token;
|
||
}
|
||
// currently implements pre-authorization
|
||
// TODO: support preAuth = false where it hooks on 401
|
||
prepareRequest(options) {
|
||
options.headers['Authorization'] = 'Bearer ' + this.token;
|
||
}
|
||
// This handler cannot handle 401
|
||
canHandleAuthentication(response) {
|
||
return false;
|
||
}
|
||
handleAuthentication(httpClient, requestInfo, objs) {
|
||
return null;
|
||
}
|
||
}
|
||
exports.BearerCredentialHandler = BearerCredentialHandler;
|
||
class PersonalAccessTokenCredentialHandler {
|
||
constructor(token) {
|
||
this.token = token;
|
||
}
|
||
// currently implements pre-authorization
|
||
// TODO: support preAuth = false where it hooks on 401
|
||
prepareRequest(options) {
|
||
options.headers['Authorization'] =
|
||
'Basic ' + Buffer.from('PAT:' + this.token).toString('base64');
|
||
}
|
||
// This handler cannot handle 401
|
||
canHandleAuthentication(response) {
|
||
return false;
|
||
}
|
||
handleAuthentication(httpClient, requestInfo, objs) {
|
||
return null;
|
||
}
|
||
}
|
||
exports.PersonalAccessTokenCredentialHandler = PersonalAccessTokenCredentialHandler;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 245:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
module.exports = globSync
|
||
globSync.GlobSync = GlobSync
|
||
|
||
var fs = __webpack_require__(747)
|
||
var rp = __webpack_require__(302)
|
||
var minimatch = __webpack_require__(93)
|
||
var Minimatch = minimatch.Minimatch
|
||
var Glob = __webpack_require__(402).Glob
|
||
var util = __webpack_require__(669)
|
||
var path = __webpack_require__(622)
|
||
var assert = __webpack_require__(357)
|
||
var isAbsolute = __webpack_require__(681)
|
||
var common = __webpack_require__(644)
|
||
var alphasort = common.alphasort
|
||
var alphasorti = common.alphasorti
|
||
var setopts = common.setopts
|
||
var ownProp = common.ownProp
|
||
var childrenIgnored = common.childrenIgnored
|
||
var isIgnored = common.isIgnored
|
||
|
||
function globSync (pattern, options) {
|
||
if (typeof options === 'function' || arguments.length === 3)
|
||
throw new TypeError('callback provided to sync glob\n'+
|
||
'See: https://github.com/isaacs/node-glob/issues/167')
|
||
|
||
return new GlobSync(pattern, options).found
|
||
}
|
||
|
||
function GlobSync (pattern, options) {
|
||
if (!pattern)
|
||
throw new Error('must provide pattern')
|
||
|
||
if (typeof options === 'function' || arguments.length === 3)
|
||
throw new TypeError('callback provided to sync glob\n'+
|
||
'See: https://github.com/isaacs/node-glob/issues/167')
|
||
|
||
if (!(this instanceof GlobSync))
|
||
return new GlobSync(pattern, options)
|
||
|
||
setopts(this, pattern, options)
|
||
|
||
if (this.noprocess)
|
||
return this
|
||
|
||
var n = this.minimatch.set.length
|
||
this.matches = new Array(n)
|
||
for (var i = 0; i < n; i ++) {
|
||
this._process(this.minimatch.set[i], i, false)
|
||
}
|
||
this._finish()
|
||
}
|
||
|
||
GlobSync.prototype._finish = function () {
|
||
assert(this instanceof GlobSync)
|
||
if (this.realpath) {
|
||
var self = this
|
||
this.matches.forEach(function (matchset, index) {
|
||
var set = self.matches[index] = Object.create(null)
|
||
for (var p in matchset) {
|
||
try {
|
||
p = self._makeAbs(p)
|
||
var real = rp.realpathSync(p, self.realpathCache)
|
||
set[real] = true
|
||
} catch (er) {
|
||
if (er.syscall === 'stat')
|
||
set[self._makeAbs(p)] = true
|
||
else
|
||
throw er
|
||
}
|
||
}
|
||
})
|
||
}
|
||
common.finish(this)
|
||
}
|
||
|
||
|
||
GlobSync.prototype._process = function (pattern, index, inGlobStar) {
|
||
assert(this instanceof GlobSync)
|
||
|
||
// Get the first [n] parts of pattern that are all strings.
|
||
var n = 0
|
||
while (typeof pattern[n] === 'string') {
|
||
n ++
|
||
}
|
||
// now n is the index of the first one that is *not* a string.
|
||
|
||
// See if there's anything else
|
||
var prefix
|
||
switch (n) {
|
||
// if not, then this is rather simple
|
||
case pattern.length:
|
||
this._processSimple(pattern.join('/'), index)
|
||
return
|
||
|
||
case 0:
|
||
// pattern *starts* with some non-trivial item.
|
||
// going to readdir(cwd), but not include the prefix in matches.
|
||
prefix = null
|
||
break
|
||
|
||
default:
|
||
// pattern has some string bits in the front.
|
||
// whatever it starts with, whether that's 'absolute' like /foo/bar,
|
||
// or 'relative' like '../baz'
|
||
prefix = pattern.slice(0, n).join('/')
|
||
break
|
||
}
|
||
|
||
var remain = pattern.slice(n)
|
||
|
||
// get the list of entries.
|
||
var read
|
||
if (prefix === null)
|
||
read = '.'
|
||
else if (isAbsolute(prefix) || isAbsolute(pattern.join('/'))) {
|
||
if (!prefix || !isAbsolute(prefix))
|
||
prefix = '/' + prefix
|
||
read = prefix
|
||
} else
|
||
read = prefix
|
||
|
||
var abs = this._makeAbs(read)
|
||
|
||
//if ignored, skip processing
|
||
if (childrenIgnored(this, read))
|
||
return
|
||
|
||
var isGlobStar = remain[0] === minimatch.GLOBSTAR
|
||
if (isGlobStar)
|
||
this._processGlobStar(prefix, read, abs, remain, index, inGlobStar)
|
||
else
|
||
this._processReaddir(prefix, read, abs, remain, index, inGlobStar)
|
||
}
|
||
|
||
|
||
GlobSync.prototype._processReaddir = function (prefix, read, abs, remain, index, inGlobStar) {
|
||
var entries = this._readdir(abs, inGlobStar)
|
||
|
||
// if the abs isn't a dir, then nothing can match!
|
||
if (!entries)
|
||
return
|
||
|
||
// It will only match dot entries if it starts with a dot, or if
|
||
// dot is set. Stuff like @(.foo|.bar) isn't allowed.
|
||
var pn = remain[0]
|
||
var negate = !!this.minimatch.negate
|
||
var rawGlob = pn._glob
|
||
var dotOk = this.dot || rawGlob.charAt(0) === '.'
|
||
|
||
var matchedEntries = []
|
||
for (var i = 0; i < entries.length; i++) {
|
||
var e = entries[i]
|
||
if (e.charAt(0) !== '.' || dotOk) {
|
||
var m
|
||
if (negate && !prefix) {
|
||
m = !e.match(pn)
|
||
} else {
|
||
m = e.match(pn)
|
||
}
|
||
if (m)
|
||
matchedEntries.push(e)
|
||
}
|
||
}
|
||
|
||
var len = matchedEntries.length
|
||
// If there are no matched entries, then nothing matches.
|
||
if (len === 0)
|
||
return
|
||
|
||
// if this is the last remaining pattern bit, then no need for
|
||
// an additional stat *unless* the user has specified mark or
|
||
// stat explicitly. We know they exist, since readdir returned
|
||
// them.
|
||
|
||
if (remain.length === 1 && !this.mark && !this.stat) {
|
||
if (!this.matches[index])
|
||
this.matches[index] = Object.create(null)
|
||
|
||
for (var i = 0; i < len; i ++) {
|
||
var e = matchedEntries[i]
|
||
if (prefix) {
|
||
if (prefix.slice(-1) !== '/')
|
||
e = prefix + '/' + e
|
||
else
|
||
e = prefix + e
|
||
}
|
||
|
||
if (e.charAt(0) === '/' && !this.nomount) {
|
||
e = path.join(this.root, e)
|
||
}
|
||
this._emitMatch(index, e)
|
||
}
|
||
// This was the last one, and no stats were needed
|
||
return
|
||
}
|
||
|
||
// now test all matched entries as stand-ins for that part
|
||
// of the pattern.
|
||
remain.shift()
|
||
for (var i = 0; i < len; i ++) {
|
||
var e = matchedEntries[i]
|
||
var newPattern
|
||
if (prefix)
|
||
newPattern = [prefix, e]
|
||
else
|
||
newPattern = [e]
|
||
this._process(newPattern.concat(remain), index, inGlobStar)
|
||
}
|
||
}
|
||
|
||
|
||
GlobSync.prototype._emitMatch = function (index, e) {
|
||
if (isIgnored(this, e))
|
||
return
|
||
|
||
var abs = this._makeAbs(e)
|
||
|
||
if (this.mark)
|
||
e = this._mark(e)
|
||
|
||
if (this.absolute) {
|
||
e = abs
|
||
}
|
||
|
||
if (this.matches[index][e])
|
||
return
|
||
|
||
if (this.nodir) {
|
||
var c = this.cache[abs]
|
||
if (c === 'DIR' || Array.isArray(c))
|
||
return
|
||
}
|
||
|
||
this.matches[index][e] = true
|
||
|
||
if (this.stat)
|
||
this._stat(e)
|
||
}
|
||
|
||
|
||
GlobSync.prototype._readdirInGlobStar = function (abs) {
|
||
// follow all symlinked directories forever
|
||
// just proceed as if this is a non-globstar situation
|
||
if (this.follow)
|
||
return this._readdir(abs, false)
|
||
|
||
var entries
|
||
var lstat
|
||
var stat
|
||
try {
|
||
lstat = fs.lstatSync(abs)
|
||
} catch (er) {
|
||
if (er.code === 'ENOENT') {
|
||
// lstat failed, doesn't exist
|
||
return null
|
||
}
|
||
}
|
||
|
||
var isSym = lstat && lstat.isSymbolicLink()
|
||
this.symlinks[abs] = isSym
|
||
|
||
// If it's not a symlink or a dir, then it's definitely a regular file.
|
||
// don't bother doing a readdir in that case.
|
||
if (!isSym && lstat && !lstat.isDirectory())
|
||
this.cache[abs] = 'FILE'
|
||
else
|
||
entries = this._readdir(abs, false)
|
||
|
||
return entries
|
||
}
|
||
|
||
GlobSync.prototype._readdir = function (abs, inGlobStar) {
|
||
var entries
|
||
|
||
if (inGlobStar && !ownProp(this.symlinks, abs))
|
||
return this._readdirInGlobStar(abs)
|
||
|
||
if (ownProp(this.cache, abs)) {
|
||
var c = this.cache[abs]
|
||
if (!c || c === 'FILE')
|
||
return null
|
||
|
||
if (Array.isArray(c))
|
||
return c
|
||
}
|
||
|
||
try {
|
||
return this._readdirEntries(abs, fs.readdirSync(abs))
|
||
} catch (er) {
|
||
this._readdirError(abs, er)
|
||
return null
|
||
}
|
||
}
|
||
|
||
GlobSync.prototype._readdirEntries = function (abs, entries) {
|
||
// if we haven't asked to stat everything, then just
|
||
// assume that everything in there exists, so we can avoid
|
||
// having to stat it a second time.
|
||
if (!this.mark && !this.stat) {
|
||
for (var i = 0; i < entries.length; i ++) {
|
||
var e = entries[i]
|
||
if (abs === '/')
|
||
e = abs + e
|
||
else
|
||
e = abs + '/' + e
|
||
this.cache[e] = true
|
||
}
|
||
}
|
||
|
||
this.cache[abs] = entries
|
||
|
||
// mark and cache dir-ness
|
||
return entries
|
||
}
|
||
|
||
GlobSync.prototype._readdirError = function (f, er) {
|
||
// handle errors, and cache the information
|
||
switch (er.code) {
|
||
case 'ENOTSUP': // https://github.com/isaacs/node-glob/issues/205
|
||
case 'ENOTDIR': // totally normal. means it *does* exist.
|
||
var abs = this._makeAbs(f)
|
||
this.cache[abs] = 'FILE'
|
||
if (abs === this.cwdAbs) {
|
||
var error = new Error(er.code + ' invalid cwd ' + this.cwd)
|
||
error.path = this.cwd
|
||
error.code = er.code
|
||
throw error
|
||
}
|
||
break
|
||
|
||
case 'ENOENT': // not terribly unusual
|
||
case 'ELOOP':
|
||
case 'ENAMETOOLONG':
|
||
case 'UNKNOWN':
|
||
this.cache[this._makeAbs(f)] = false
|
||
break
|
||
|
||
default: // some unusual error. Treat as failure.
|
||
this.cache[this._makeAbs(f)] = false
|
||
if (this.strict)
|
||
throw er
|
||
if (!this.silent)
|
||
console.error('glob error', er)
|
||
break
|
||
}
|
||
}
|
||
|
||
GlobSync.prototype._processGlobStar = function (prefix, read, abs, remain, index, inGlobStar) {
|
||
|
||
var entries = this._readdir(abs, inGlobStar)
|
||
|
||
// no entries means not a dir, so it can never have matches
|
||
// foo.txt/** doesn't match foo.txt
|
||
if (!entries)
|
||
return
|
||
|
||
// test without the globstar, and with every child both below
|
||
// and replacing the globstar.
|
||
var remainWithoutGlobStar = remain.slice(1)
|
||
var gspref = prefix ? [ prefix ] : []
|
||
var noGlobStar = gspref.concat(remainWithoutGlobStar)
|
||
|
||
// the noGlobStar pattern exits the inGlobStar state
|
||
this._process(noGlobStar, index, false)
|
||
|
||
var len = entries.length
|
||
var isSym = this.symlinks[abs]
|
||
|
||
// If it's a symlink, and we're in a globstar, then stop
|
||
if (isSym && inGlobStar)
|
||
return
|
||
|
||
for (var i = 0; i < len; i++) {
|
||
var e = entries[i]
|
||
if (e.charAt(0) === '.' && !this.dot)
|
||
continue
|
||
|
||
// these two cases enter the inGlobStar state
|
||
var instead = gspref.concat(entries[i], remainWithoutGlobStar)
|
||
this._process(instead, index, true)
|
||
|
||
var below = gspref.concat(entries[i], remain)
|
||
this._process(below, index, true)
|
||
}
|
||
}
|
||
|
||
GlobSync.prototype._processSimple = function (prefix, index) {
|
||
// XXX review this. Shouldn't it be doing the mounting etc
|
||
// before doing stat? kinda weird?
|
||
var exists = this._stat(prefix)
|
||
|
||
if (!this.matches[index])
|
||
this.matches[index] = Object.create(null)
|
||
|
||
// If it doesn't exist, then just mark the lack of results
|
||
if (!exists)
|
||
return
|
||
|
||
if (prefix && isAbsolute(prefix) && !this.nomount) {
|
||
var trail = /[\/\\]$/.test(prefix)
|
||
if (prefix.charAt(0) === '/') {
|
||
prefix = path.join(this.root, prefix)
|
||
} else {
|
||
prefix = path.resolve(this.root, prefix)
|
||
if (trail)
|
||
prefix += '/'
|
||
}
|
||
}
|
||
|
||
if (process.platform === 'win32')
|
||
prefix = prefix.replace(/\\/g, '/')
|
||
|
||
// Mark this as a match
|
||
this._emitMatch(index, prefix)
|
||
}
|
||
|
||
// Returns either 'DIR', 'FILE', or false
|
||
GlobSync.prototype._stat = function (f) {
|
||
var abs = this._makeAbs(f)
|
||
var needDir = f.slice(-1) === '/'
|
||
|
||
if (f.length > this.maxLength)
|
||
return false
|
||
|
||
if (!this.stat && ownProp(this.cache, abs)) {
|
||
var c = this.cache[abs]
|
||
|
||
if (Array.isArray(c))
|
||
c = 'DIR'
|
||
|
||
// It exists, but maybe not how we need it
|
||
if (!needDir || c === 'DIR')
|
||
return c
|
||
|
||
if (needDir && c === 'FILE')
|
||
return false
|
||
|
||
// otherwise we have to stat, because maybe c=true
|
||
// if we know it exists, but not what it is.
|
||
}
|
||
|
||
var exists
|
||
var stat = this.statCache[abs]
|
||
if (!stat) {
|
||
var lstat
|
||
try {
|
||
lstat = fs.lstatSync(abs)
|
||
} catch (er) {
|
||
if (er && (er.code === 'ENOENT' || er.code === 'ENOTDIR')) {
|
||
this.statCache[abs] = false
|
||
return false
|
||
}
|
||
}
|
||
|
||
if (lstat && lstat.isSymbolicLink()) {
|
||
try {
|
||
stat = fs.statSync(abs)
|
||
} catch (er) {
|
||
stat = lstat
|
||
}
|
||
} else {
|
||
stat = lstat
|
||
}
|
||
}
|
||
|
||
this.statCache[abs] = stat
|
||
|
||
var c = true
|
||
if (stat)
|
||
c = stat.isDirectory() ? 'DIR' : 'FILE'
|
||
|
||
this.cache[abs] = this.cache[abs] || c
|
||
|
||
if (needDir && c === 'FILE')
|
||
return false
|
||
|
||
return c
|
||
}
|
||
|
||
GlobSync.prototype._mark = function (p) {
|
||
return common.mark(this, p)
|
||
}
|
||
|
||
GlobSync.prototype._makeAbs = function (f) {
|
||
return common.makeAbs(this, f)
|
||
}
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 247:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
const os = __webpack_require__(87);
|
||
const tty = __webpack_require__(867);
|
||
const hasFlag = __webpack_require__(364);
|
||
|
||
const {env} = process;
|
||
|
||
let forceColor;
|
||
if (hasFlag('no-color') ||
|
||
hasFlag('no-colors') ||
|
||
hasFlag('color=false') ||
|
||
hasFlag('color=never')) {
|
||
forceColor = 0;
|
||
} else if (hasFlag('color') ||
|
||
hasFlag('colors') ||
|
||
hasFlag('color=true') ||
|
||
hasFlag('color=always')) {
|
||
forceColor = 1;
|
||
}
|
||
|
||
if ('FORCE_COLOR' in env) {
|
||
if (env.FORCE_COLOR === 'true') {
|
||
forceColor = 1;
|
||
} else if (env.FORCE_COLOR === 'false') {
|
||
forceColor = 0;
|
||
} else {
|
||
forceColor = env.FORCE_COLOR.length === 0 ? 1 : Math.min(parseInt(env.FORCE_COLOR, 10), 3);
|
||
}
|
||
}
|
||
|
||
function translateLevel(level) {
|
||
if (level === 0) {
|
||
return false;
|
||
}
|
||
|
||
return {
|
||
level,
|
||
hasBasic: true,
|
||
has256: level >= 2,
|
||
has16m: level >= 3
|
||
};
|
||
}
|
||
|
||
function supportsColor(haveStream, streamIsTTY) {
|
||
if (forceColor === 0) {
|
||
return 0;
|
||
}
|
||
|
||
if (hasFlag('color=16m') ||
|
||
hasFlag('color=full') ||
|
||
hasFlag('color=truecolor')) {
|
||
return 3;
|
||
}
|
||
|
||
if (hasFlag('color=256')) {
|
||
return 2;
|
||
}
|
||
|
||
if (haveStream && !streamIsTTY && forceColor === undefined) {
|
||
return 0;
|
||
}
|
||
|
||
const min = forceColor || 0;
|
||
|
||
if (env.TERM === 'dumb') {
|
||
return min;
|
||
}
|
||
|
||
if (process.platform === 'win32') {
|
||
// Windows 10 build 10586 is the first Windows release that supports 256 colors.
|
||
// Windows 10 build 14931 is the first release that supports 16m/TrueColor.
|
||
const osRelease = os.release().split('.');
|
||
if (
|
||
Number(osRelease[0]) >= 10 &&
|
||
Number(osRelease[2]) >= 10586
|
||
) {
|
||
return Number(osRelease[2]) >= 14931 ? 3 : 2;
|
||
}
|
||
|
||
return 1;
|
||
}
|
||
|
||
if ('CI' in env) {
|
||
if (['TRAVIS', 'CIRCLECI', 'APPVEYOR', 'GITLAB_CI'].some(sign => sign in env) || env.CI_NAME === 'codeship') {
|
||
return 1;
|
||
}
|
||
|
||
return min;
|
||
}
|
||
|
||
if ('TEAMCITY_VERSION' in env) {
|
||
return /^(9\.(0*[1-9]\d*)\.|\d{2,}\.)/.test(env.TEAMCITY_VERSION) ? 1 : 0;
|
||
}
|
||
|
||
if ('GITHUB_ACTIONS' in env) {
|
||
return 1;
|
||
}
|
||
|
||
if (env.COLORTERM === 'truecolor') {
|
||
return 3;
|
||
}
|
||
|
||
if ('TERM_PROGRAM' in env) {
|
||
const version = parseInt((env.TERM_PROGRAM_VERSION || '').split('.')[0], 10);
|
||
|
||
switch (env.TERM_PROGRAM) {
|
||
case 'iTerm.app':
|
||
return version >= 3 ? 3 : 2;
|
||
case 'Apple_Terminal':
|
||
return 2;
|
||
// No default
|
||
}
|
||
}
|
||
|
||
if (/-256(color)?$/i.test(env.TERM)) {
|
||
return 2;
|
||
}
|
||
|
||
if (/^screen|^xterm|^vt100|^vt220|^rxvt|color|ansi|cygwin|linux/i.test(env.TERM)) {
|
||
return 1;
|
||
}
|
||
|
||
if ('COLORTERM' in env) {
|
||
return 1;
|
||
}
|
||
|
||
return min;
|
||
}
|
||
|
||
function getSupportLevel(stream) {
|
||
const level = supportsColor(stream, stream && stream.isTTY);
|
||
return translateLevel(level);
|
||
}
|
||
|
||
module.exports = {
|
||
supportsColor: getSupportLevel,
|
||
stdout: translateLevel(supportsColor(true, tty.isatty(1))),
|
||
stderr: translateLevel(supportsColor(true, tty.isatty(2)))
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 281:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
||
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
||
return new (P || (P = Promise))(function (resolve, reject) {
|
||
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
||
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
||
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
||
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
||
});
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const internal_globber_1 = __webpack_require__(297);
|
||
/**
|
||
* Constructs a globber
|
||
*
|
||
* @param patterns Patterns separated by newlines
|
||
* @param options Glob options
|
||
*/
|
||
function create(patterns, options) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
return yield internal_globber_1.DefaultGlobber.create(patterns, options);
|
||
});
|
||
}
|
||
exports.create = create;
|
||
//# sourceMappingURL=glob.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 283:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var utils = __webpack_require__(35);
|
||
|
||
function InterceptorManager() {
|
||
this.handlers = [];
|
||
}
|
||
|
||
/**
|
||
* Add a new interceptor to the stack
|
||
*
|
||
* @param {Function} fulfilled The function to handle `then` for a `Promise`
|
||
* @param {Function} rejected The function to handle `reject` for a `Promise`
|
||
*
|
||
* @return {Number} An ID used to remove interceptor later
|
||
*/
|
||
InterceptorManager.prototype.use = function use(fulfilled, rejected, options) {
|
||
this.handlers.push({
|
||
fulfilled: fulfilled,
|
||
rejected: rejected,
|
||
synchronous: options ? options.synchronous : false,
|
||
runWhen: options ? options.runWhen : null
|
||
});
|
||
return this.handlers.length - 1;
|
||
};
|
||
|
||
/**
|
||
* Remove an interceptor from the stack
|
||
*
|
||
* @param {Number} id The ID that was returned by `use`
|
||
*/
|
||
InterceptorManager.prototype.eject = function eject(id) {
|
||
if (this.handlers[id]) {
|
||
this.handlers[id] = null;
|
||
}
|
||
};
|
||
|
||
/**
|
||
* Iterate over all the registered interceptors
|
||
*
|
||
* This method is particularly useful for skipping over any
|
||
* interceptors that may have become `null` calling `eject`.
|
||
*
|
||
* @param {Function} fn The function to call for each interceptor
|
||
*/
|
||
InterceptorManager.prototype.forEach = function forEach(fn) {
|
||
utils.forEach(this.handlers, function forEachHandler(h) {
|
||
if (h !== null) {
|
||
fn(h);
|
||
}
|
||
});
|
||
};
|
||
|
||
module.exports = InterceptorManager;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 297:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
||
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
||
return new (P || (P = Promise))(function (resolve, reject) {
|
||
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
||
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
||
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
||
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
||
});
|
||
};
|
||
var __asyncValues = (this && this.__asyncValues) || function (o) {
|
||
if (!Symbol.asyncIterator) throw new TypeError("Symbol.asyncIterator is not defined.");
|
||
var m = o[Symbol.asyncIterator], i;
|
||
return m ? m.call(o) : (o = typeof __values === "function" ? __values(o) : o[Symbol.iterator](), i = {}, verb("next"), verb("throw"), verb("return"), i[Symbol.asyncIterator] = function () { return this; }, i);
|
||
function verb(n) { i[n] = o[n] && function (v) { return new Promise(function (resolve, reject) { v = o[n](v), settle(resolve, reject, v.done, v.value); }); }; }
|
||
function settle(resolve, reject, d, v) { Promise.resolve(v).then(function(v) { resolve({ value: v, done: d }); }, reject); }
|
||
};
|
||
var __await = (this && this.__await) || function (v) { return this instanceof __await ? (this.v = v, this) : new __await(v); }
|
||
var __asyncGenerator = (this && this.__asyncGenerator) || function (thisArg, _arguments, generator) {
|
||
if (!Symbol.asyncIterator) throw new TypeError("Symbol.asyncIterator is not defined.");
|
||
var g = generator.apply(thisArg, _arguments || []), i, q = [];
|
||
return i = {}, verb("next"), verb("throw"), verb("return"), i[Symbol.asyncIterator] = function () { return this; }, i;
|
||
function verb(n) { if (g[n]) i[n] = function (v) { return new Promise(function (a, b) { q.push([n, v, a, b]) > 1 || resume(n, v); }); }; }
|
||
function resume(n, v) { try { step(g[n](v)); } catch (e) { settle(q[0][3], e); } }
|
||
function step(r) { r.value instanceof __await ? Promise.resolve(r.value.v).then(fulfill, reject) : settle(q[0][2], r); }
|
||
function fulfill(value) { resume("next", value); }
|
||
function reject(value) { resume("throw", value); }
|
||
function settle(f, v) { if (f(v), q.shift(), q.length) resume(q[0][0], q[0][1]); }
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const core = __webpack_require__(470);
|
||
const fs = __webpack_require__(747);
|
||
const globOptionsHelper = __webpack_require__(601);
|
||
const path = __webpack_require__(622);
|
||
const patternHelper = __webpack_require__(597);
|
||
const internal_match_kind_1 = __webpack_require__(327);
|
||
const internal_pattern_1 = __webpack_require__(923);
|
||
const internal_search_state_1 = __webpack_require__(728);
|
||
const IS_WINDOWS = process.platform === 'win32';
|
||
class DefaultGlobber {
|
||
constructor(options) {
|
||
this.patterns = [];
|
||
this.searchPaths = [];
|
||
this.options = globOptionsHelper.getOptions(options);
|
||
}
|
||
getSearchPaths() {
|
||
// Return a copy
|
||
return this.searchPaths.slice();
|
||
}
|
||
glob() {
|
||
var e_1, _a;
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
const result = [];
|
||
try {
|
||
for (var _b = __asyncValues(this.globGenerator()), _c; _c = yield _b.next(), !_c.done;) {
|
||
const itemPath = _c.value;
|
||
result.push(itemPath);
|
||
}
|
||
}
|
||
catch (e_1_1) { e_1 = { error: e_1_1 }; }
|
||
finally {
|
||
try {
|
||
if (_c && !_c.done && (_a = _b.return)) yield _a.call(_b);
|
||
}
|
||
finally { if (e_1) throw e_1.error; }
|
||
}
|
||
return result;
|
||
});
|
||
}
|
||
globGenerator() {
|
||
return __asyncGenerator(this, arguments, function* globGenerator_1() {
|
||
// Fill in defaults options
|
||
const options = globOptionsHelper.getOptions(this.options);
|
||
// Implicit descendants?
|
||
const patterns = [];
|
||
for (const pattern of this.patterns) {
|
||
patterns.push(pattern);
|
||
if (options.implicitDescendants &&
|
||
(pattern.trailingSeparator ||
|
||
pattern.segments[pattern.segments.length - 1] !== '**')) {
|
||
patterns.push(new internal_pattern_1.Pattern(pattern.negate, pattern.segments.concat('**')));
|
||
}
|
||
}
|
||
// Push the search paths
|
||
const stack = [];
|
||
for (const searchPath of patternHelper.getSearchPaths(patterns)) {
|
||
core.debug(`Search path '${searchPath}'`);
|
||
// Exists?
|
||
try {
|
||
// Intentionally using lstat. Detection for broken symlink
|
||
// will be performed later (if following symlinks).
|
||
yield __await(fs.promises.lstat(searchPath));
|
||
}
|
||
catch (err) {
|
||
if (err.code === 'ENOENT') {
|
||
continue;
|
||
}
|
||
throw err;
|
||
}
|
||
stack.unshift(new internal_search_state_1.SearchState(searchPath, 1));
|
||
}
|
||
// Search
|
||
const traversalChain = []; // used to detect cycles
|
||
while (stack.length) {
|
||
// Pop
|
||
const item = stack.pop();
|
||
// Match?
|
||
const match = patternHelper.match(patterns, item.path);
|
||
const partialMatch = !!match || patternHelper.partialMatch(patterns, item.path);
|
||
if (!match && !partialMatch) {
|
||
continue;
|
||
}
|
||
// Stat
|
||
const stats = yield __await(DefaultGlobber.stat(item, options, traversalChain)
|
||
// Broken symlink, or symlink cycle detected, or no longer exists
|
||
);
|
||
// Broken symlink, or symlink cycle detected, or no longer exists
|
||
if (!stats) {
|
||
continue;
|
||
}
|
||
// Directory
|
||
if (stats.isDirectory()) {
|
||
// Matched
|
||
if (match & internal_match_kind_1.MatchKind.Directory) {
|
||
yield yield __await(item.path);
|
||
}
|
||
// Descend?
|
||
else if (!partialMatch) {
|
||
continue;
|
||
}
|
||
// Push the child items in reverse
|
||
const childLevel = item.level + 1;
|
||
const childItems = (yield __await(fs.promises.readdir(item.path))).map(x => new internal_search_state_1.SearchState(path.join(item.path, x), childLevel));
|
||
stack.push(...childItems.reverse());
|
||
}
|
||
// File
|
||
else if (match & internal_match_kind_1.MatchKind.File) {
|
||
yield yield __await(item.path);
|
||
}
|
||
}
|
||
});
|
||
}
|
||
/**
|
||
* Constructs a DefaultGlobber
|
||
*/
|
||
static create(patterns, options) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
const result = new DefaultGlobber(options);
|
||
if (IS_WINDOWS) {
|
||
patterns = patterns.replace(/\r\n/g, '\n');
|
||
patterns = patterns.replace(/\r/g, '\n');
|
||
}
|
||
const lines = patterns.split('\n').map(x => x.trim());
|
||
for (const line of lines) {
|
||
// Empty or comment
|
||
if (!line || line.startsWith('#')) {
|
||
continue;
|
||
}
|
||
// Pattern
|
||
else {
|
||
result.patterns.push(new internal_pattern_1.Pattern(line));
|
||
}
|
||
}
|
||
result.searchPaths.push(...patternHelper.getSearchPaths(result.patterns));
|
||
return result;
|
||
});
|
||
}
|
||
static stat(item, options, traversalChain) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
// Note:
|
||
// `stat` returns info about the target of a symlink (or symlink chain)
|
||
// `lstat` returns info about a symlink itself
|
||
let stats;
|
||
if (options.followSymbolicLinks) {
|
||
try {
|
||
// Use `stat` (following symlinks)
|
||
stats = yield fs.promises.stat(item.path);
|
||
}
|
||
catch (err) {
|
||
if (err.code === 'ENOENT') {
|
||
if (options.omitBrokenSymbolicLinks) {
|
||
core.debug(`Broken symlink '${item.path}'`);
|
||
return undefined;
|
||
}
|
||
throw new Error(`No information found for the path '${item.path}'. This may indicate a broken symbolic link.`);
|
||
}
|
||
throw err;
|
||
}
|
||
}
|
||
else {
|
||
// Use `lstat` (not following symlinks)
|
||
stats = yield fs.promises.lstat(item.path);
|
||
}
|
||
// Note, isDirectory() returns false for the lstat of a symlink
|
||
if (stats.isDirectory() && options.followSymbolicLinks) {
|
||
// Get the realpath
|
||
const realPath = yield fs.promises.realpath(item.path);
|
||
// Fixup the traversal chain to match the item level
|
||
while (traversalChain.length >= item.level) {
|
||
traversalChain.pop();
|
||
}
|
||
// Test for a cycle
|
||
if (traversalChain.some((x) => x === realPath)) {
|
||
core.debug(`Symlink cycle detected for path '${item.path}' and realpath '${realPath}'`);
|
||
return undefined;
|
||
}
|
||
// Update the traversal chain
|
||
traversalChain.push(realPath);
|
||
}
|
||
return stats;
|
||
});
|
||
}
|
||
}
|
||
exports.DefaultGlobber = DefaultGlobber;
|
||
//# sourceMappingURL=internal-globber.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 302:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
module.exports = realpath
|
||
realpath.realpath = realpath
|
||
realpath.sync = realpathSync
|
||
realpath.realpathSync = realpathSync
|
||
realpath.monkeypatch = monkeypatch
|
||
realpath.unmonkeypatch = unmonkeypatch
|
||
|
||
var fs = __webpack_require__(747)
|
||
var origRealpath = fs.realpath
|
||
var origRealpathSync = fs.realpathSync
|
||
|
||
var version = process.version
|
||
var ok = /^v[0-5]\./.test(version)
|
||
var old = __webpack_require__(117)
|
||
|
||
function newError (er) {
|
||
return er && er.syscall === 'realpath' && (
|
||
er.code === 'ELOOP' ||
|
||
er.code === 'ENOMEM' ||
|
||
er.code === 'ENAMETOOLONG'
|
||
)
|
||
}
|
||
|
||
function realpath (p, cache, cb) {
|
||
if (ok) {
|
||
return origRealpath(p, cache, cb)
|
||
}
|
||
|
||
if (typeof cache === 'function') {
|
||
cb = cache
|
||
cache = null
|
||
}
|
||
origRealpath(p, cache, function (er, result) {
|
||
if (newError(er)) {
|
||
old.realpath(p, cache, cb)
|
||
} else {
|
||
cb(er, result)
|
||
}
|
||
})
|
||
}
|
||
|
||
function realpathSync (p, cache) {
|
||
if (ok) {
|
||
return origRealpathSync(p, cache)
|
||
}
|
||
|
||
try {
|
||
return origRealpathSync(p, cache)
|
||
} catch (er) {
|
||
if (newError(er)) {
|
||
return old.realpathSync(p, cache)
|
||
} else {
|
||
throw er
|
||
}
|
||
}
|
||
}
|
||
|
||
function monkeypatch () {
|
||
fs.realpath = realpath
|
||
fs.realpathSync = realpathSync
|
||
}
|
||
|
||
function unmonkeypatch () {
|
||
fs.realpath = origRealpath
|
||
fs.realpathSync = origRealpathSync
|
||
}
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 306:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
var concatMap = __webpack_require__(896);
|
||
var balanced = __webpack_require__(621);
|
||
|
||
module.exports = expandTop;
|
||
|
||
var escSlash = '\0SLASH'+Math.random()+'\0';
|
||
var escOpen = '\0OPEN'+Math.random()+'\0';
|
||
var escClose = '\0CLOSE'+Math.random()+'\0';
|
||
var escComma = '\0COMMA'+Math.random()+'\0';
|
||
var escPeriod = '\0PERIOD'+Math.random()+'\0';
|
||
|
||
function numeric(str) {
|
||
return parseInt(str, 10) == str
|
||
? parseInt(str, 10)
|
||
: str.charCodeAt(0);
|
||
}
|
||
|
||
function escapeBraces(str) {
|
||
return str.split('\\\\').join(escSlash)
|
||
.split('\\{').join(escOpen)
|
||
.split('\\}').join(escClose)
|
||
.split('\\,').join(escComma)
|
||
.split('\\.').join(escPeriod);
|
||
}
|
||
|
||
function unescapeBraces(str) {
|
||
return str.split(escSlash).join('\\')
|
||
.split(escOpen).join('{')
|
||
.split(escClose).join('}')
|
||
.split(escComma).join(',')
|
||
.split(escPeriod).join('.');
|
||
}
|
||
|
||
|
||
// Basically just str.split(","), but handling cases
|
||
// where we have nested braced sections, which should be
|
||
// treated as individual members, like {a,{b,c},d}
|
||
function parseCommaParts(str) {
|
||
if (!str)
|
||
return [''];
|
||
|
||
var parts = [];
|
||
var m = balanced('{', '}', str);
|
||
|
||
if (!m)
|
||
return str.split(',');
|
||
|
||
var pre = m.pre;
|
||
var body = m.body;
|
||
var post = m.post;
|
||
var p = pre.split(',');
|
||
|
||
p[p.length-1] += '{' + body + '}';
|
||
var postParts = parseCommaParts(post);
|
||
if (post.length) {
|
||
p[p.length-1] += postParts.shift();
|
||
p.push.apply(p, postParts);
|
||
}
|
||
|
||
parts.push.apply(parts, p);
|
||
|
||
return parts;
|
||
}
|
||
|
||
function expandTop(str) {
|
||
if (!str)
|
||
return [];
|
||
|
||
// I don't know why Bash 4.3 does this, but it does.
|
||
// Anything starting with {} will have the first two bytes preserved
|
||
// but *only* at the top level, so {},a}b will not expand to anything,
|
||
// but a{},b}c will be expanded to [a}c,abc].
|
||
// One could argue that this is a bug in Bash, but since the goal of
|
||
// this module is to match Bash's rules, we escape a leading {}
|
||
if (str.substr(0, 2) === '{}') {
|
||
str = '\\{\\}' + str.substr(2);
|
||
}
|
||
|
||
return expand(escapeBraces(str), true).map(unescapeBraces);
|
||
}
|
||
|
||
function identity(e) {
|
||
return e;
|
||
}
|
||
|
||
function embrace(str) {
|
||
return '{' + str + '}';
|
||
}
|
||
function isPadded(el) {
|
||
return /^-?0\d/.test(el);
|
||
}
|
||
|
||
function lte(i, y) {
|
||
return i <= y;
|
||
}
|
||
function gte(i, y) {
|
||
return i >= y;
|
||
}
|
||
|
||
function expand(str, isTop) {
|
||
var expansions = [];
|
||
|
||
var m = balanced('{', '}', str);
|
||
if (!m || /\$$/.test(m.pre)) return [str];
|
||
|
||
var isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body);
|
||
var isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body);
|
||
var isSequence = isNumericSequence || isAlphaSequence;
|
||
var isOptions = m.body.indexOf(',') >= 0;
|
||
if (!isSequence && !isOptions) {
|
||
// {a},b}
|
||
if (m.post.match(/,.*\}/)) {
|
||
str = m.pre + '{' + m.body + escClose + m.post;
|
||
return expand(str);
|
||
}
|
||
return [str];
|
||
}
|
||
|
||
var n;
|
||
if (isSequence) {
|
||
n = m.body.split(/\.\./);
|
||
} else {
|
||
n = parseCommaParts(m.body);
|
||
if (n.length === 1) {
|
||
// x{{a,b}}y ==> x{a}y x{b}y
|
||
n = expand(n[0], false).map(embrace);
|
||
if (n.length === 1) {
|
||
var post = m.post.length
|
||
? expand(m.post, false)
|
||
: [''];
|
||
return post.map(function(p) {
|
||
return m.pre + n[0] + p;
|
||
});
|
||
}
|
||
}
|
||
}
|
||
|
||
// at this point, n is the parts, and we know it's not a comma set
|
||
// with a single entry.
|
||
|
||
// no need to expand pre, since it is guaranteed to be free of brace-sets
|
||
var pre = m.pre;
|
||
var post = m.post.length
|
||
? expand(m.post, false)
|
||
: [''];
|
||
|
||
var N;
|
||
|
||
if (isSequence) {
|
||
var x = numeric(n[0]);
|
||
var y = numeric(n[1]);
|
||
var width = Math.max(n[0].length, n[1].length)
|
||
var incr = n.length == 3
|
||
? Math.abs(numeric(n[2]))
|
||
: 1;
|
||
var test = lte;
|
||
var reverse = y < x;
|
||
if (reverse) {
|
||
incr *= -1;
|
||
test = gte;
|
||
}
|
||
var pad = n.some(isPadded);
|
||
|
||
N = [];
|
||
|
||
for (var i = x; test(i, y); i += incr) {
|
||
var c;
|
||
if (isAlphaSequence) {
|
||
c = String.fromCharCode(i);
|
||
if (c === '\\')
|
||
c = '';
|
||
} else {
|
||
c = String(i);
|
||
if (pad) {
|
||
var need = width - c.length;
|
||
if (need > 0) {
|
||
var z = new Array(need + 1).join('0');
|
||
if (i < 0)
|
||
c = '-' + z + c.slice(1);
|
||
else
|
||
c = z + c;
|
||
}
|
||
}
|
||
}
|
||
N.push(c);
|
||
}
|
||
} else {
|
||
N = concatMap(n, function(el) { return expand(el, false) });
|
||
}
|
||
|
||
for (var j = 0; j < N.length; j++) {
|
||
for (var k = 0; k < post.length; k++) {
|
||
var expansion = pre + N[j] + post[k];
|
||
if (!isTop || isSequence || expansion)
|
||
expansions.push(expansion);
|
||
}
|
||
}
|
||
|
||
return expansions;
|
||
}
|
||
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 315:
|
||
/***/ (function(module) {
|
||
|
||
if (typeof Object.create === 'function') {
|
||
// implementation from standard node.js 'util' module
|
||
module.exports = function inherits(ctor, superCtor) {
|
||
if (superCtor) {
|
||
ctor.super_ = superCtor
|
||
ctor.prototype = Object.create(superCtor.prototype, {
|
||
constructor: {
|
||
value: ctor,
|
||
enumerable: false,
|
||
writable: true,
|
||
configurable: true
|
||
}
|
||
})
|
||
}
|
||
};
|
||
} else {
|
||
// old school shim for old browsers
|
||
module.exports = function inherits(ctor, superCtor) {
|
||
if (superCtor) {
|
||
ctor.super_ = superCtor
|
||
var TempCtor = function () {}
|
||
TempCtor.prototype = superCtor.prototype
|
||
ctor.prototype = new TempCtor()
|
||
ctor.prototype.constructor = ctor
|
||
}
|
||
}
|
||
}
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 327:
|
||
/***/ (function(__unusedmodule, exports) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
/**
|
||
* Indicates whether a pattern matches a path
|
||
*/
|
||
var MatchKind;
|
||
(function (MatchKind) {
|
||
/** Not matched */
|
||
MatchKind[MatchKind["None"] = 0] = "None";
|
||
/** Matched if the path is a directory */
|
||
MatchKind[MatchKind["Directory"] = 1] = "Directory";
|
||
/** Matched if the path is a regular file */
|
||
MatchKind[MatchKind["File"] = 2] = "File";
|
||
/** Matched */
|
||
MatchKind[MatchKind["All"] = 3] = "All";
|
||
})(MatchKind = exports.MatchKind || (exports.MatchKind = {}));
|
||
//# sourceMappingURL=internal-match-kind.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 352:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var utils = __webpack_require__(35);
|
||
var bind = __webpack_require__(727);
|
||
var Axios = __webpack_require__(779);
|
||
var mergeConfig = __webpack_require__(825);
|
||
var defaults = __webpack_require__(529);
|
||
|
||
/**
|
||
* Create an instance of Axios
|
||
*
|
||
* @param {Object} defaultConfig The default config for the instance
|
||
* @return {Axios} A new instance of Axios
|
||
*/
|
||
function createInstance(defaultConfig) {
|
||
var context = new Axios(defaultConfig);
|
||
var instance = bind(Axios.prototype.request, context);
|
||
|
||
// Copy axios.prototype to instance
|
||
utils.extend(instance, Axios.prototype, context);
|
||
|
||
// Copy context to instance
|
||
utils.extend(instance, context);
|
||
|
||
return instance;
|
||
}
|
||
|
||
// Create the default instance to be exported
|
||
var axios = createInstance(defaults);
|
||
|
||
// Expose Axios class to allow class inheritance
|
||
axios.Axios = Axios;
|
||
|
||
// Factory for creating new instances
|
||
axios.create = function create(instanceConfig) {
|
||
return createInstance(mergeConfig(axios.defaults, instanceConfig));
|
||
};
|
||
|
||
// Expose Cancel & CancelToken
|
||
axios.Cancel = __webpack_require__(826);
|
||
axios.CancelToken = __webpack_require__(137);
|
||
axios.isCancel = __webpack_require__(732);
|
||
|
||
// Expose all/spread
|
||
axios.all = function all(promises) {
|
||
return Promise.all(promises);
|
||
};
|
||
axios.spread = __webpack_require__(879);
|
||
|
||
// Expose isAxiosError
|
||
axios.isAxiosError = __webpack_require__(104);
|
||
|
||
module.exports = axios;
|
||
|
||
// Allow use of default import syntax in TypeScript
|
||
module.exports.default = axios;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 357:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("assert");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 359:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
||
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
||
return new (P || (P = Promise))(function (resolve, reject) {
|
||
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
||
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
||
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
||
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
||
});
|
||
};
|
||
var __importStar = (this && this.__importStar) || function (mod) {
|
||
if (mod && mod.__esModule) return mod;
|
||
var result = {};
|
||
if (mod != null) for (var k in mod) if (Object.hasOwnProperty.call(mod, k)) result[k] = mod[k];
|
||
result["default"] = mod;
|
||
return result;
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const core = __importStar(__webpack_require__(470));
|
||
const upload_specification_1 = __webpack_require__(862);
|
||
const upload_http_client_1 = __webpack_require__(608);
|
||
const utils_1 = __webpack_require__(870);
|
||
const download_http_client_1 = __webpack_require__(855);
|
||
const download_specification_1 = __webpack_require__(532);
|
||
const config_variables_1 = __webpack_require__(401);
|
||
const path_1 = __webpack_require__(622);
|
||
class DefaultArtifactClient {
|
||
/**
|
||
* Constructs a DefaultArtifactClient
|
||
*/
|
||
static create() {
|
||
return new DefaultArtifactClient();
|
||
}
|
||
/**
|
||
* Uploads an artifact
|
||
*/
|
||
uploadArtifact(name, files, rootDirectory, options) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
utils_1.checkArtifactName(name);
|
||
// Get specification for the files being uploaded
|
||
const uploadSpecification = upload_specification_1.getUploadSpecification(name, rootDirectory, files);
|
||
const uploadResponse = {
|
||
artifactName: name,
|
||
artifactItems: [],
|
||
size: 0,
|
||
failedItems: []
|
||
};
|
||
const uploadHttpClient = new upload_http_client_1.UploadHttpClient();
|
||
if (uploadSpecification.length === 0) {
|
||
core.warning(`No files found that can be uploaded`);
|
||
}
|
||
else {
|
||
// Create an entry for the artifact in the file container
|
||
const response = yield uploadHttpClient.createArtifactInFileContainer(name, options);
|
||
if (!response.fileContainerResourceUrl) {
|
||
core.debug(response.toString());
|
||
throw new Error('No URL provided by the Artifact Service to upload an artifact to');
|
||
}
|
||
core.debug(`Upload Resource URL: ${response.fileContainerResourceUrl}`);
|
||
// Upload each of the files that were found concurrently
|
||
const uploadResult = yield uploadHttpClient.uploadArtifactToFileContainer(response.fileContainerResourceUrl, uploadSpecification, options);
|
||
// Update the size of the artifact to indicate we are done uploading
|
||
// The uncompressed size is used for display when downloading a zip of the artifact from the UI
|
||
yield uploadHttpClient.patchArtifactSize(uploadResult.totalSize, name);
|
||
core.info(`Finished uploading artifact ${name}. Reported size is ${uploadResult.uploadSize} bytes. There were ${uploadResult.failedItems.length} items that failed to upload`);
|
||
uploadResponse.artifactItems = uploadSpecification.map(item => item.absoluteFilePath);
|
||
uploadResponse.size = uploadResult.uploadSize;
|
||
uploadResponse.failedItems = uploadResult.failedItems;
|
||
}
|
||
return uploadResponse;
|
||
});
|
||
}
|
||
downloadArtifact(name, path, options) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
const downloadHttpClient = new download_http_client_1.DownloadHttpClient();
|
||
const artifacts = yield downloadHttpClient.listArtifacts();
|
||
if (artifacts.count === 0) {
|
||
throw new Error(`Unable to find any artifacts for the associated workflow`);
|
||
}
|
||
const artifactToDownload = artifacts.value.find(artifact => {
|
||
return artifact.name === name;
|
||
});
|
||
if (!artifactToDownload) {
|
||
throw new Error(`Unable to find an artifact with the name: ${name}`);
|
||
}
|
||
const items = yield downloadHttpClient.getContainerItems(artifactToDownload.name, artifactToDownload.fileContainerResourceUrl);
|
||
if (!path) {
|
||
path = config_variables_1.getWorkSpaceDirectory();
|
||
}
|
||
path = path_1.normalize(path);
|
||
path = path_1.resolve(path);
|
||
// During upload, empty directories are rejected by the remote server so there should be no artifacts that consist of only empty directories
|
||
const downloadSpecification = download_specification_1.getDownloadSpecification(name, items.value, path, (options === null || options === void 0 ? void 0 : options.createArtifactFolder) || false);
|
||
if (downloadSpecification.filesToDownload.length === 0) {
|
||
core.info(`No downloadable files were found for the artifact: ${artifactToDownload.name}`);
|
||
}
|
||
else {
|
||
// Create all necessary directories recursively before starting any download
|
||
yield utils_1.createDirectoriesForArtifact(downloadSpecification.directoryStructure);
|
||
core.info('Directory structure has been setup for the artifact');
|
||
yield utils_1.createEmptyFilesForArtifact(downloadSpecification.emptyFilesToCreate);
|
||
yield downloadHttpClient.downloadSingleArtifact(downloadSpecification.filesToDownload);
|
||
}
|
||
return {
|
||
artifactName: name,
|
||
downloadPath: downloadSpecification.rootDownloadLocation
|
||
};
|
||
});
|
||
}
|
||
downloadAllArtifacts(path) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
const downloadHttpClient = new download_http_client_1.DownloadHttpClient();
|
||
const response = [];
|
||
const artifacts = yield downloadHttpClient.listArtifacts();
|
||
if (artifacts.count === 0) {
|
||
core.info('Unable to find any artifacts for the associated workflow');
|
||
return response;
|
||
}
|
||
if (!path) {
|
||
path = config_variables_1.getWorkSpaceDirectory();
|
||
}
|
||
path = path_1.normalize(path);
|
||
path = path_1.resolve(path);
|
||
let downloadedArtifacts = 0;
|
||
while (downloadedArtifacts < artifacts.count) {
|
||
const currentArtifactToDownload = artifacts.value[downloadedArtifacts];
|
||
downloadedArtifacts += 1;
|
||
// Get container entries for the specific artifact
|
||
const items = yield downloadHttpClient.getContainerItems(currentArtifactToDownload.name, currentArtifactToDownload.fileContainerResourceUrl);
|
||
const downloadSpecification = download_specification_1.getDownloadSpecification(currentArtifactToDownload.name, items.value, path, true);
|
||
if (downloadSpecification.filesToDownload.length === 0) {
|
||
core.info(`No downloadable files were found for any artifact ${currentArtifactToDownload.name}`);
|
||
}
|
||
else {
|
||
yield utils_1.createDirectoriesForArtifact(downloadSpecification.directoryStructure);
|
||
yield utils_1.createEmptyFilesForArtifact(downloadSpecification.emptyFilesToCreate);
|
||
yield downloadHttpClient.downloadSingleArtifact(downloadSpecification.filesToDownload);
|
||
}
|
||
response.push({
|
||
artifactName: currentArtifactToDownload.name,
|
||
downloadPath: downloadSpecification.rootDownloadLocation
|
||
});
|
||
}
|
||
return response;
|
||
});
|
||
}
|
||
}
|
||
exports.DefaultArtifactClient = DefaultArtifactClient;
|
||
//# sourceMappingURL=artifact-client.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 361:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = {"_from":"axios","_id":"axios@0.21.4","_inBundle":false,"_integrity":"sha512-ut5vewkiu8jjGBdqpM44XxjuCjq9LAKeHVmoVfHVzy8eHgxxq8SbAVQNovDA8mVi05kP0Ea/n/UzcSHcTJQfNg==","_location":"/axios","_phantomChildren":{},"_requested":{"type":"tag","registry":true,"raw":"axios","name":"axios","escapedName":"axios","rawSpec":"","saveSpec":null,"fetchSpec":"latest"},"_requiredBy":["#USER","/"],"_resolved":"https://registry.npmjs.org/axios/-/axios-0.21.4.tgz","_shasum":"c67b90dc0568e5c1cf2b0b858c43ba28e2eda575","_spec":"axios","_where":"/Users/mingzi/github/upload-artifact","author":{"name":"Matt Zabriskie"},"browser":{"./lib/adapters/http.js":"./lib/adapters/xhr.js"},"bugs":{"url":"https://github.com/axios/axios/issues"},"bundleDependencies":false,"bundlesize":[{"path":"./dist/axios.min.js","threshold":"5kB"}],"dependencies":{"follow-redirects":"^1.14.0"},"deprecated":false,"description":"Promise based HTTP client for the browser and node.js","devDependencies":{"coveralls":"^3.0.0","es6-promise":"^4.2.4","grunt":"^1.3.0","grunt-banner":"^0.6.0","grunt-cli":"^1.2.0","grunt-contrib-clean":"^1.1.0","grunt-contrib-watch":"^1.0.0","grunt-eslint":"^23.0.0","grunt-karma":"^4.0.0","grunt-mocha-test":"^0.13.3","grunt-ts":"^6.0.0-beta.19","grunt-webpack":"^4.0.2","istanbul-instrumenter-loader":"^1.0.0","jasmine-core":"^2.4.1","karma":"^6.3.2","karma-chrome-launcher":"^3.1.0","karma-firefox-launcher":"^2.1.0","karma-jasmine":"^1.1.1","karma-jasmine-ajax":"^0.1.13","karma-safari-launcher":"^1.0.0","karma-sauce-launcher":"^4.3.6","karma-sinon":"^1.0.5","karma-sourcemap-loader":"^0.3.8","karma-webpack":"^4.0.2","load-grunt-tasks":"^3.5.2","minimist":"^1.2.0","mocha":"^8.2.1","sinon":"^4.5.0","terser-webpack-plugin":"^4.2.3","typescript":"^4.0.5","url-search-params":"^0.10.0","webpack":"^4.44.2","webpack-dev-server":"^3.11.0"},"homepage":"https://axios-http.com","jsdelivr":"dist/axios.min.js","keywords":["xhr","http","ajax","promise","node"],"license":"MIT","main":"index.js","name":"axios","repository":{"type":"git","url":"git+https://github.com/axios/axios.git"},"scripts":{"build":"NODE_ENV=production grunt build","coveralls":"cat coverage/lcov.info | ./node_modules/coveralls/bin/coveralls.js","examples":"node ./examples/server.js","fix":"eslint --fix lib/**/*.js","postversion":"git push && git push --tags","preversion":"npm test","start":"node ./sandbox/server.js","test":"grunt test","version":"npm run build && grunt version && git add -A dist && git add CHANGELOG.md bower.json package.json"},"typings":"./index.d.ts","unpkg":"dist/axios.min.js","version":"0.21.4"};
|
||
|
||
/***/ }),
|
||
|
||
/***/ 364:
|
||
/***/ (function(module) {
|
||
|
||
"use strict";
|
||
|
||
|
||
module.exports = (flag, argv = process.argv) => {
|
||
const prefix = flag.startsWith('-') ? '' : (flag.length === 1 ? '-' : '--');
|
||
const position = argv.indexOf(prefix + flag);
|
||
const terminatorPosition = argv.indexOf('--');
|
||
return position !== -1 && (terminatorPosition === -1 || position < terminatorPosition);
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 369:
|
||
/***/ (function(module) {
|
||
|
||
"use strict";
|
||
|
||
|
||
/**
|
||
* Update an Error with the specified config, error code, and response.
|
||
*
|
||
* @param {Error} error The error to update.
|
||
* @param {Object} config The config.
|
||
* @param {string} [code] The error code (for example, 'ECONNABORTED').
|
||
* @param {Object} [request] The request.
|
||
* @param {Object} [response] The response.
|
||
* @returns {Error} The error.
|
||
*/
|
||
module.exports = function enhanceError(error, config, code, request, response) {
|
||
error.config = config;
|
||
if (code) {
|
||
error.code = code;
|
||
}
|
||
|
||
error.request = request;
|
||
error.response = response;
|
||
error.isAxiosError = true;
|
||
|
||
error.toJSON = function toJSON() {
|
||
return {
|
||
// Standard
|
||
message: this.message,
|
||
name: this.name,
|
||
// Microsoft
|
||
description: this.description,
|
||
number: this.number,
|
||
// Mozilla
|
||
fileName: this.fileName,
|
||
lineNumber: this.lineNumber,
|
||
columnNumber: this.columnNumber,
|
||
stack: this.stack,
|
||
// Axios
|
||
config: this.config,
|
||
code: this.code
|
||
};
|
||
};
|
||
return error;
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 383:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const assert = __webpack_require__(357);
|
||
const path = __webpack_require__(622);
|
||
const pathHelper = __webpack_require__(972);
|
||
const IS_WINDOWS = process.platform === 'win32';
|
||
/**
|
||
* Helper class for parsing paths into segments
|
||
*/
|
||
class Path {
|
||
/**
|
||
* Constructs a Path
|
||
* @param itemPath Path or array of segments
|
||
*/
|
||
constructor(itemPath) {
|
||
this.segments = [];
|
||
// String
|
||
if (typeof itemPath === 'string') {
|
||
assert(itemPath, `Parameter 'itemPath' must not be empty`);
|
||
// Normalize slashes and trim unnecessary trailing slash
|
||
itemPath = pathHelper.safeTrimTrailingSeparator(itemPath);
|
||
// Not rooted
|
||
if (!pathHelper.hasRoot(itemPath)) {
|
||
this.segments = itemPath.split(path.sep);
|
||
}
|
||
// Rooted
|
||
else {
|
||
// Add all segments, while not at the root
|
||
let remaining = itemPath;
|
||
let dir = pathHelper.dirname(remaining);
|
||
while (dir !== remaining) {
|
||
// Add the segment
|
||
const basename = path.basename(remaining);
|
||
this.segments.unshift(basename);
|
||
// Truncate the last segment
|
||
remaining = dir;
|
||
dir = pathHelper.dirname(remaining);
|
||
}
|
||
// Remainder is the root
|
||
this.segments.unshift(remaining);
|
||
}
|
||
}
|
||
// Array
|
||
else {
|
||
// Must not be empty
|
||
assert(itemPath.length > 0, `Parameter 'itemPath' must not be an empty array`);
|
||
// Each segment
|
||
for (let i = 0; i < itemPath.length; i++) {
|
||
let segment = itemPath[i];
|
||
// Must not be empty
|
||
assert(segment, `Parameter 'itemPath' must not contain any empty segments`);
|
||
// Normalize slashes
|
||
segment = pathHelper.normalizeSeparators(itemPath[i]);
|
||
// Root segment
|
||
if (i === 0 && pathHelper.hasRoot(segment)) {
|
||
segment = pathHelper.safeTrimTrailingSeparator(segment);
|
||
assert(segment === pathHelper.dirname(segment), `Parameter 'itemPath' root segment contains information for multiple segments`);
|
||
this.segments.push(segment);
|
||
}
|
||
// All other segments
|
||
else {
|
||
// Must not contain slash
|
||
assert(!segment.includes(path.sep), `Parameter 'itemPath' contains unexpected path separators`);
|
||
this.segments.push(segment);
|
||
}
|
||
}
|
||
}
|
||
}
|
||
/**
|
||
* Converts the path to it's string representation
|
||
*/
|
||
toString() {
|
||
// First segment
|
||
let result = this.segments[0];
|
||
// All others
|
||
let skipSlash = result.endsWith(path.sep) || (IS_WINDOWS && /^[A-Z]:$/i.test(result));
|
||
for (let i = 1; i < this.segments.length; i++) {
|
||
if (skipSlash) {
|
||
skipSlash = false;
|
||
}
|
||
else {
|
||
result += path.sep;
|
||
}
|
||
result += this.segments[i];
|
||
}
|
||
return result;
|
||
}
|
||
}
|
||
exports.Path = Path;
|
||
//# sourceMappingURL=internal-path.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 385:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
||
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
||
return new (P || (P = Promise))(function (resolve, reject) {
|
||
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
||
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
||
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
||
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
||
});
|
||
};
|
||
var __importStar = (this && this.__importStar) || function (mod) {
|
||
if (mod && mod.__esModule) return mod;
|
||
var result = {};
|
||
if (mod != null) for (var k in mod) if (Object.hasOwnProperty.call(mod, k)) result[k] = mod[k];
|
||
result["default"] = mod;
|
||
return result;
|
||
};
|
||
var __importDefault = (this && this.__importDefault) || function (mod) {
|
||
return (mod && mod.__esModule) ? mod : { "default": mod };
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const core = __importStar(__webpack_require__(470));
|
||
const axios_1 = __importDefault(__webpack_require__(53));
|
||
const artifact_1 = __webpack_require__(214);
|
||
const search_1 = __webpack_require__(575);
|
||
const input_helper_1 = __webpack_require__(583);
|
||
const constants_1 = __webpack_require__(694);
|
||
function run() {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
try {
|
||
const inputs = input_helper_1.getInputs();
|
||
const searchResult = yield search_1.findFilesToUpload(inputs.searchPath);
|
||
if (searchResult.filesToUpload.length === 0) {
|
||
// No files were found, different use cases warrant different types of behavior if nothing is found
|
||
switch (inputs.ifNoFilesFound) {
|
||
case constants_1.NoFileOptions.warn: {
|
||
core.warning(`No files were found with the provided path: ${inputs.searchPath}. No artifacts will be uploaded.`);
|
||
break;
|
||
}
|
||
case constants_1.NoFileOptions.error: {
|
||
core.setFailed(`No files were found with the provided path: ${inputs.searchPath}. No artifacts will be uploaded.`);
|
||
break;
|
||
}
|
||
case constants_1.NoFileOptions.ignore: {
|
||
core.info(`No files were found with the provided path: ${inputs.searchPath}. No artifacts will be uploaded.`);
|
||
break;
|
||
}
|
||
}
|
||
}
|
||
else {
|
||
const s = searchResult.filesToUpload.length === 1 ? '' : 's';
|
||
core.info(`With the provided path, there will be ${searchResult.filesToUpload.length} file${s} uploaded`);
|
||
core.debug(`Root artifact directory is ${searchResult.rootDirectory}`);
|
||
if (searchResult.filesToUpload.length > 10000) {
|
||
core.warning(`There are over 10,000 files in this artifact, consider creating an archive before upload to improve the upload performance.`);
|
||
}
|
||
const artifactClient = artifact_1.create();
|
||
const options = {
|
||
continueOnError: false
|
||
};
|
||
if (inputs.retentionDays) {
|
||
options.retentionDays = inputs.retentionDays;
|
||
}
|
||
const uploadResponse = yield artifactClient.uploadArtifact(inputs.artifactName, searchResult.filesToUpload, searchResult.rootDirectory, options);
|
||
if (uploadResponse.failedItems.length > 0) {
|
||
core.setFailed(`An error was encountered when uploading ${uploadResponse.artifactName}. There were ${uploadResponse.failedItems.length} items that failed to upload.`);
|
||
}
|
||
else {
|
||
core.info(`Artifact ${uploadResponse.artifactName} has been successfully uploaded!`);
|
||
let runtimeUrl = process.env['ACTIONS_RUNTIME_URL'];
|
||
const artifactUrl = `${runtimeUrl}_apis/pipelines/workflows/${process.env['GITHUB_RUN_ID']}/artifacts?api-version=6.0-preview`;
|
||
let response = yield axios_1.default.get(artifactUrl, {
|
||
headers: {
|
||
"Authorization": `Bearer ${process.env["ACTIONS_RUNTIME_TOKEN"]}`,
|
||
"Content-Type": "application/json"
|
||
}
|
||
});
|
||
const unsignedUrl = `${response.data.value[0].url}`;
|
||
console.log(response.data);
|
||
console.log(`unsigned artifact url is ${unsignedUrl}`);
|
||
if (!inputs.token) {
|
||
console.log("missing github token");
|
||
}
|
||
else {
|
||
console.log(`https://api.github.com/repos/${process.env['GITHUB_REPOSITORY']}/pages/deployment`);
|
||
console.log(`token: ${inputs.token}`);
|
||
response = yield axios_1.default.post(`https://api.github.com/repos/${process.env['GITHUB_REPOSITORY']}/pages/deployment`, {
|
||
artifact_url: unsignedUrl,
|
||
pages_build_version: process.env['GITHUB_SHA']
|
||
}, {
|
||
headers: {
|
||
"Accept": "application/vnd.github.v3+json",
|
||
"Content-Type": "application/json",
|
||
"Authorization": `Bearer ${inputs.token}}`
|
||
}
|
||
});
|
||
console.log(response.data);
|
||
}
|
||
}
|
||
}
|
||
}
|
||
catch (err) {
|
||
core.setFailed(err.message);
|
||
}
|
||
});
|
||
}
|
||
run();
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 401:
|
||
/***/ (function(__unusedmodule, exports) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
// The number of concurrent uploads that happens at the same time
|
||
function getUploadFileConcurrency() {
|
||
return 2;
|
||
}
|
||
exports.getUploadFileConcurrency = getUploadFileConcurrency;
|
||
// When uploading large files that can't be uploaded with a single http call, this controls
|
||
// the chunk size that is used during upload
|
||
function getUploadChunkSize() {
|
||
return 8 * 1024 * 1024; // 8 MB Chunks
|
||
}
|
||
exports.getUploadChunkSize = getUploadChunkSize;
|
||
// The maximum number of retries that can be attempted before an upload or download fails
|
||
function getRetryLimit() {
|
||
return 5;
|
||
}
|
||
exports.getRetryLimit = getRetryLimit;
|
||
// With exponential backoff, the larger the retry count, the larger the wait time before another attempt
|
||
// The retry multiplier controls by how much the backOff time increases depending on the number of retries
|
||
function getRetryMultiplier() {
|
||
return 1.5;
|
||
}
|
||
exports.getRetryMultiplier = getRetryMultiplier;
|
||
// The initial wait time if an upload or download fails and a retry is being attempted for the first time
|
||
function getInitialRetryIntervalInMilliseconds() {
|
||
return 3000;
|
||
}
|
||
exports.getInitialRetryIntervalInMilliseconds = getInitialRetryIntervalInMilliseconds;
|
||
// The number of concurrent downloads that happens at the same time
|
||
function getDownloadFileConcurrency() {
|
||
return 2;
|
||
}
|
||
exports.getDownloadFileConcurrency = getDownloadFileConcurrency;
|
||
function getRuntimeToken() {
|
||
const token = process.env['ACTIONS_RUNTIME_TOKEN'];
|
||
if (!token) {
|
||
throw new Error('Unable to get ACTIONS_RUNTIME_TOKEN env variable');
|
||
}
|
||
return token;
|
||
}
|
||
exports.getRuntimeToken = getRuntimeToken;
|
||
function getRuntimeUrl() {
|
||
const runtimeUrl = process.env['ACTIONS_RUNTIME_URL'];
|
||
if (!runtimeUrl) {
|
||
throw new Error('Unable to get ACTIONS_RUNTIME_URL env variable');
|
||
}
|
||
return runtimeUrl;
|
||
}
|
||
exports.getRuntimeUrl = getRuntimeUrl;
|
||
function getWorkFlowRunId() {
|
||
const workFlowRunId = process.env['GITHUB_RUN_ID'];
|
||
if (!workFlowRunId) {
|
||
throw new Error('Unable to get GITHUB_RUN_ID env variable');
|
||
}
|
||
return workFlowRunId;
|
||
}
|
||
exports.getWorkFlowRunId = getWorkFlowRunId;
|
||
function getWorkSpaceDirectory() {
|
||
const workspaceDirectory = process.env['GITHUB_WORKSPACE'];
|
||
if (!workspaceDirectory) {
|
||
throw new Error('Unable to get GITHUB_WORKSPACE env variable');
|
||
}
|
||
return workspaceDirectory;
|
||
}
|
||
exports.getWorkSpaceDirectory = getWorkSpaceDirectory;
|
||
function getRetentionDays() {
|
||
return process.env['GITHUB_RETENTION_DAYS'];
|
||
}
|
||
exports.getRetentionDays = getRetentionDays;
|
||
//# sourceMappingURL=config-variables.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 402:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
// Approach:
|
||
//
|
||
// 1. Get the minimatch set
|
||
// 2. For each pattern in the set, PROCESS(pattern, false)
|
||
// 3. Store matches per-set, then uniq them
|
||
//
|
||
// PROCESS(pattern, inGlobStar)
|
||
// Get the first [n] items from pattern that are all strings
|
||
// Join these together. This is PREFIX.
|
||
// If there is no more remaining, then stat(PREFIX) and
|
||
// add to matches if it succeeds. END.
|
||
//
|
||
// If inGlobStar and PREFIX is symlink and points to dir
|
||
// set ENTRIES = []
|
||
// else readdir(PREFIX) as ENTRIES
|
||
// If fail, END
|
||
//
|
||
// with ENTRIES
|
||
// If pattern[n] is GLOBSTAR
|
||
// // handle the case where the globstar match is empty
|
||
// // by pruning it out, and testing the resulting pattern
|
||
// PROCESS(pattern[0..n] + pattern[n+1 .. $], false)
|
||
// // handle other cases.
|
||
// for ENTRY in ENTRIES (not dotfiles)
|
||
// // attach globstar + tail onto the entry
|
||
// // Mark that this entry is a globstar match
|
||
// PROCESS(pattern[0..n] + ENTRY + pattern[n .. $], true)
|
||
//
|
||
// else // not globstar
|
||
// for ENTRY in ENTRIES (not dotfiles, unless pattern[n] is dot)
|
||
// Test ENTRY against pattern[n]
|
||
// If fails, continue
|
||
// If passes, PROCESS(pattern[0..n] + item + pattern[n+1 .. $])
|
||
//
|
||
// Caveat:
|
||
// Cache all stats and readdirs results to minimize syscall. Since all
|
||
// we ever care about is existence and directory-ness, we can just keep
|
||
// `true` for files, and [children,...] for directories, or `false` for
|
||
// things that don't exist.
|
||
|
||
module.exports = glob
|
||
|
||
var fs = __webpack_require__(747)
|
||
var rp = __webpack_require__(302)
|
||
var minimatch = __webpack_require__(93)
|
||
var Minimatch = minimatch.Minimatch
|
||
var inherits = __webpack_require__(689)
|
||
var EE = __webpack_require__(614).EventEmitter
|
||
var path = __webpack_require__(622)
|
||
var assert = __webpack_require__(357)
|
||
var isAbsolute = __webpack_require__(681)
|
||
var globSync = __webpack_require__(245)
|
||
var common = __webpack_require__(644)
|
||
var alphasort = common.alphasort
|
||
var alphasorti = common.alphasorti
|
||
var setopts = common.setopts
|
||
var ownProp = common.ownProp
|
||
var inflight = __webpack_require__(674)
|
||
var util = __webpack_require__(669)
|
||
var childrenIgnored = common.childrenIgnored
|
||
var isIgnored = common.isIgnored
|
||
|
||
var once = __webpack_require__(49)
|
||
|
||
function glob (pattern, options, cb) {
|
||
if (typeof options === 'function') cb = options, options = {}
|
||
if (!options) options = {}
|
||
|
||
if (options.sync) {
|
||
if (cb)
|
||
throw new TypeError('callback provided to sync glob')
|
||
return globSync(pattern, options)
|
||
}
|
||
|
||
return new Glob(pattern, options, cb)
|
||
}
|
||
|
||
glob.sync = globSync
|
||
var GlobSync = glob.GlobSync = globSync.GlobSync
|
||
|
||
// old api surface
|
||
glob.glob = glob
|
||
|
||
function extend (origin, add) {
|
||
if (add === null || typeof add !== 'object') {
|
||
return origin
|
||
}
|
||
|
||
var keys = Object.keys(add)
|
||
var i = keys.length
|
||
while (i--) {
|
||
origin[keys[i]] = add[keys[i]]
|
||
}
|
||
return origin
|
||
}
|
||
|
||
glob.hasMagic = function (pattern, options_) {
|
||
var options = extend({}, options_)
|
||
options.noprocess = true
|
||
|
||
var g = new Glob(pattern, options)
|
||
var set = g.minimatch.set
|
||
|
||
if (!pattern)
|
||
return false
|
||
|
||
if (set.length > 1)
|
||
return true
|
||
|
||
for (var j = 0; j < set[0].length; j++) {
|
||
if (typeof set[0][j] !== 'string')
|
||
return true
|
||
}
|
||
|
||
return false
|
||
}
|
||
|
||
glob.Glob = Glob
|
||
inherits(Glob, EE)
|
||
function Glob (pattern, options, cb) {
|
||
if (typeof options === 'function') {
|
||
cb = options
|
||
options = null
|
||
}
|
||
|
||
if (options && options.sync) {
|
||
if (cb)
|
||
throw new TypeError('callback provided to sync glob')
|
||
return new GlobSync(pattern, options)
|
||
}
|
||
|
||
if (!(this instanceof Glob))
|
||
return new Glob(pattern, options, cb)
|
||
|
||
setopts(this, pattern, options)
|
||
this._didRealPath = false
|
||
|
||
// process each pattern in the minimatch set
|
||
var n = this.minimatch.set.length
|
||
|
||
// The matches are stored as {<filename>: true,...} so that
|
||
// duplicates are automagically pruned.
|
||
// Later, we do an Object.keys() on these.
|
||
// Keep them as a list so we can fill in when nonull is set.
|
||
this.matches = new Array(n)
|
||
|
||
if (typeof cb === 'function') {
|
||
cb = once(cb)
|
||
this.on('error', cb)
|
||
this.on('end', function (matches) {
|
||
cb(null, matches)
|
||
})
|
||
}
|
||
|
||
var self = this
|
||
this._processing = 0
|
||
|
||
this._emitQueue = []
|
||
this._processQueue = []
|
||
this.paused = false
|
||
|
||
if (this.noprocess)
|
||
return this
|
||
|
||
if (n === 0)
|
||
return done()
|
||
|
||
var sync = true
|
||
for (var i = 0; i < n; i ++) {
|
||
this._process(this.minimatch.set[i], i, false, done)
|
||
}
|
||
sync = false
|
||
|
||
function done () {
|
||
--self._processing
|
||
if (self._processing <= 0) {
|
||
if (sync) {
|
||
process.nextTick(function () {
|
||
self._finish()
|
||
})
|
||
} else {
|
||
self._finish()
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
Glob.prototype._finish = function () {
|
||
assert(this instanceof Glob)
|
||
if (this.aborted)
|
||
return
|
||
|
||
if (this.realpath && !this._didRealpath)
|
||
return this._realpath()
|
||
|
||
common.finish(this)
|
||
this.emit('end', this.found)
|
||
}
|
||
|
||
Glob.prototype._realpath = function () {
|
||
if (this._didRealpath)
|
||
return
|
||
|
||
this._didRealpath = true
|
||
|
||
var n = this.matches.length
|
||
if (n === 0)
|
||
return this._finish()
|
||
|
||
var self = this
|
||
for (var i = 0; i < this.matches.length; i++)
|
||
this._realpathSet(i, next)
|
||
|
||
function next () {
|
||
if (--n === 0)
|
||
self._finish()
|
||
}
|
||
}
|
||
|
||
Glob.prototype._realpathSet = function (index, cb) {
|
||
var matchset = this.matches[index]
|
||
if (!matchset)
|
||
return cb()
|
||
|
||
var found = Object.keys(matchset)
|
||
var self = this
|
||
var n = found.length
|
||
|
||
if (n === 0)
|
||
return cb()
|
||
|
||
var set = this.matches[index] = Object.create(null)
|
||
found.forEach(function (p, i) {
|
||
// If there's a problem with the stat, then it means that
|
||
// one or more of the links in the realpath couldn't be
|
||
// resolved. just return the abs value in that case.
|
||
p = self._makeAbs(p)
|
||
rp.realpath(p, self.realpathCache, function (er, real) {
|
||
if (!er)
|
||
set[real] = true
|
||
else if (er.syscall === 'stat')
|
||
set[p] = true
|
||
else
|
||
self.emit('error', er) // srsly wtf right here
|
||
|
||
if (--n === 0) {
|
||
self.matches[index] = set
|
||
cb()
|
||
}
|
||
})
|
||
})
|
||
}
|
||
|
||
Glob.prototype._mark = function (p) {
|
||
return common.mark(this, p)
|
||
}
|
||
|
||
Glob.prototype._makeAbs = function (f) {
|
||
return common.makeAbs(this, f)
|
||
}
|
||
|
||
Glob.prototype.abort = function () {
|
||
this.aborted = true
|
||
this.emit('abort')
|
||
}
|
||
|
||
Glob.prototype.pause = function () {
|
||
if (!this.paused) {
|
||
this.paused = true
|
||
this.emit('pause')
|
||
}
|
||
}
|
||
|
||
Glob.prototype.resume = function () {
|
||
if (this.paused) {
|
||
this.emit('resume')
|
||
this.paused = false
|
||
if (this._emitQueue.length) {
|
||
var eq = this._emitQueue.slice(0)
|
||
this._emitQueue.length = 0
|
||
for (var i = 0; i < eq.length; i ++) {
|
||
var e = eq[i]
|
||
this._emitMatch(e[0], e[1])
|
||
}
|
||
}
|
||
if (this._processQueue.length) {
|
||
var pq = this._processQueue.slice(0)
|
||
this._processQueue.length = 0
|
||
for (var i = 0; i < pq.length; i ++) {
|
||
var p = pq[i]
|
||
this._processing--
|
||
this._process(p[0], p[1], p[2], p[3])
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
Glob.prototype._process = function (pattern, index, inGlobStar, cb) {
|
||
assert(this instanceof Glob)
|
||
assert(typeof cb === 'function')
|
||
|
||
if (this.aborted)
|
||
return
|
||
|
||
this._processing++
|
||
if (this.paused) {
|
||
this._processQueue.push([pattern, index, inGlobStar, cb])
|
||
return
|
||
}
|
||
|
||
//console.error('PROCESS %d', this._processing, pattern)
|
||
|
||
// Get the first [n] parts of pattern that are all strings.
|
||
var n = 0
|
||
while (typeof pattern[n] === 'string') {
|
||
n ++
|
||
}
|
||
// now n is the index of the first one that is *not* a string.
|
||
|
||
// see if there's anything else
|
||
var prefix
|
||
switch (n) {
|
||
// if not, then this is rather simple
|
||
case pattern.length:
|
||
this._processSimple(pattern.join('/'), index, cb)
|
||
return
|
||
|
||
case 0:
|
||
// pattern *starts* with some non-trivial item.
|
||
// going to readdir(cwd), but not include the prefix in matches.
|
||
prefix = null
|
||
break
|
||
|
||
default:
|
||
// pattern has some string bits in the front.
|
||
// whatever it starts with, whether that's 'absolute' like /foo/bar,
|
||
// or 'relative' like '../baz'
|
||
prefix = pattern.slice(0, n).join('/')
|
||
break
|
||
}
|
||
|
||
var remain = pattern.slice(n)
|
||
|
||
// get the list of entries.
|
||
var read
|
||
if (prefix === null)
|
||
read = '.'
|
||
else if (isAbsolute(prefix) || isAbsolute(pattern.join('/'))) {
|
||
if (!prefix || !isAbsolute(prefix))
|
||
prefix = '/' + prefix
|
||
read = prefix
|
||
} else
|
||
read = prefix
|
||
|
||
var abs = this._makeAbs(read)
|
||
|
||
//if ignored, skip _processing
|
||
if (childrenIgnored(this, read))
|
||
return cb()
|
||
|
||
var isGlobStar = remain[0] === minimatch.GLOBSTAR
|
||
if (isGlobStar)
|
||
this._processGlobStar(prefix, read, abs, remain, index, inGlobStar, cb)
|
||
else
|
||
this._processReaddir(prefix, read, abs, remain, index, inGlobStar, cb)
|
||
}
|
||
|
||
Glob.prototype._processReaddir = function (prefix, read, abs, remain, index, inGlobStar, cb) {
|
||
var self = this
|
||
this._readdir(abs, inGlobStar, function (er, entries) {
|
||
return self._processReaddir2(prefix, read, abs, remain, index, inGlobStar, entries, cb)
|
||
})
|
||
}
|
||
|
||
Glob.prototype._processReaddir2 = function (prefix, read, abs, remain, index, inGlobStar, entries, cb) {
|
||
|
||
// if the abs isn't a dir, then nothing can match!
|
||
if (!entries)
|
||
return cb()
|
||
|
||
// It will only match dot entries if it starts with a dot, or if
|
||
// dot is set. Stuff like @(.foo|.bar) isn't allowed.
|
||
var pn = remain[0]
|
||
var negate = !!this.minimatch.negate
|
||
var rawGlob = pn._glob
|
||
var dotOk = this.dot || rawGlob.charAt(0) === '.'
|
||
|
||
var matchedEntries = []
|
||
for (var i = 0; i < entries.length; i++) {
|
||
var e = entries[i]
|
||
if (e.charAt(0) !== '.' || dotOk) {
|
||
var m
|
||
if (negate && !prefix) {
|
||
m = !e.match(pn)
|
||
} else {
|
||
m = e.match(pn)
|
||
}
|
||
if (m)
|
||
matchedEntries.push(e)
|
||
}
|
||
}
|
||
|
||
//console.error('prd2', prefix, entries, remain[0]._glob, matchedEntries)
|
||
|
||
var len = matchedEntries.length
|
||
// If there are no matched entries, then nothing matches.
|
||
if (len === 0)
|
||
return cb()
|
||
|
||
// if this is the last remaining pattern bit, then no need for
|
||
// an additional stat *unless* the user has specified mark or
|
||
// stat explicitly. We know they exist, since readdir returned
|
||
// them.
|
||
|
||
if (remain.length === 1 && !this.mark && !this.stat) {
|
||
if (!this.matches[index])
|
||
this.matches[index] = Object.create(null)
|
||
|
||
for (var i = 0; i < len; i ++) {
|
||
var e = matchedEntries[i]
|
||
if (prefix) {
|
||
if (prefix !== '/')
|
||
e = prefix + '/' + e
|
||
else
|
||
e = prefix + e
|
||
}
|
||
|
||
if (e.charAt(0) === '/' && !this.nomount) {
|
||
e = path.join(this.root, e)
|
||
}
|
||
this._emitMatch(index, e)
|
||
}
|
||
// This was the last one, and no stats were needed
|
||
return cb()
|
||
}
|
||
|
||
// now test all matched entries as stand-ins for that part
|
||
// of the pattern.
|
||
remain.shift()
|
||
for (var i = 0; i < len; i ++) {
|
||
var e = matchedEntries[i]
|
||
var newPattern
|
||
if (prefix) {
|
||
if (prefix !== '/')
|
||
e = prefix + '/' + e
|
||
else
|
||
e = prefix + e
|
||
}
|
||
this._process([e].concat(remain), index, inGlobStar, cb)
|
||
}
|
||
cb()
|
||
}
|
||
|
||
Glob.prototype._emitMatch = function (index, e) {
|
||
if (this.aborted)
|
||
return
|
||
|
||
if (isIgnored(this, e))
|
||
return
|
||
|
||
if (this.paused) {
|
||
this._emitQueue.push([index, e])
|
||
return
|
||
}
|
||
|
||
var abs = isAbsolute(e) ? e : this._makeAbs(e)
|
||
|
||
if (this.mark)
|
||
e = this._mark(e)
|
||
|
||
if (this.absolute)
|
||
e = abs
|
||
|
||
if (this.matches[index][e])
|
||
return
|
||
|
||
if (this.nodir) {
|
||
var c = this.cache[abs]
|
||
if (c === 'DIR' || Array.isArray(c))
|
||
return
|
||
}
|
||
|
||
this.matches[index][e] = true
|
||
|
||
var st = this.statCache[abs]
|
||
if (st)
|
||
this.emit('stat', e, st)
|
||
|
||
this.emit('match', e)
|
||
}
|
||
|
||
Glob.prototype._readdirInGlobStar = function (abs, cb) {
|
||
if (this.aborted)
|
||
return
|
||
|
||
// follow all symlinked directories forever
|
||
// just proceed as if this is a non-globstar situation
|
||
if (this.follow)
|
||
return this._readdir(abs, false, cb)
|
||
|
||
var lstatkey = 'lstat\0' + abs
|
||
var self = this
|
||
var lstatcb = inflight(lstatkey, lstatcb_)
|
||
|
||
if (lstatcb)
|
||
fs.lstat(abs, lstatcb)
|
||
|
||
function lstatcb_ (er, lstat) {
|
||
if (er && er.code === 'ENOENT')
|
||
return cb()
|
||
|
||
var isSym = lstat && lstat.isSymbolicLink()
|
||
self.symlinks[abs] = isSym
|
||
|
||
// If it's not a symlink or a dir, then it's definitely a regular file.
|
||
// don't bother doing a readdir in that case.
|
||
if (!isSym && lstat && !lstat.isDirectory()) {
|
||
self.cache[abs] = 'FILE'
|
||
cb()
|
||
} else
|
||
self._readdir(abs, false, cb)
|
||
}
|
||
}
|
||
|
||
Glob.prototype._readdir = function (abs, inGlobStar, cb) {
|
||
if (this.aborted)
|
||
return
|
||
|
||
cb = inflight('readdir\0'+abs+'\0'+inGlobStar, cb)
|
||
if (!cb)
|
||
return
|
||
|
||
//console.error('RD %j %j', +inGlobStar, abs)
|
||
if (inGlobStar && !ownProp(this.symlinks, abs))
|
||
return this._readdirInGlobStar(abs, cb)
|
||
|
||
if (ownProp(this.cache, abs)) {
|
||
var c = this.cache[abs]
|
||
if (!c || c === 'FILE')
|
||
return cb()
|
||
|
||
if (Array.isArray(c))
|
||
return cb(null, c)
|
||
}
|
||
|
||
var self = this
|
||
fs.readdir(abs, readdirCb(this, abs, cb))
|
||
}
|
||
|
||
function readdirCb (self, abs, cb) {
|
||
return function (er, entries) {
|
||
if (er)
|
||
self._readdirError(abs, er, cb)
|
||
else
|
||
self._readdirEntries(abs, entries, cb)
|
||
}
|
||
}
|
||
|
||
Glob.prototype._readdirEntries = function (abs, entries, cb) {
|
||
if (this.aborted)
|
||
return
|
||
|
||
// if we haven't asked to stat everything, then just
|
||
// assume that everything in there exists, so we can avoid
|
||
// having to stat it a second time.
|
||
if (!this.mark && !this.stat) {
|
||
for (var i = 0; i < entries.length; i ++) {
|
||
var e = entries[i]
|
||
if (abs === '/')
|
||
e = abs + e
|
||
else
|
||
e = abs + '/' + e
|
||
this.cache[e] = true
|
||
}
|
||
}
|
||
|
||
this.cache[abs] = entries
|
||
return cb(null, entries)
|
||
}
|
||
|
||
Glob.prototype._readdirError = function (f, er, cb) {
|
||
if (this.aborted)
|
||
return
|
||
|
||
// handle errors, and cache the information
|
||
switch (er.code) {
|
||
case 'ENOTSUP': // https://github.com/isaacs/node-glob/issues/205
|
||
case 'ENOTDIR': // totally normal. means it *does* exist.
|
||
var abs = this._makeAbs(f)
|
||
this.cache[abs] = 'FILE'
|
||
if (abs === this.cwdAbs) {
|
||
var error = new Error(er.code + ' invalid cwd ' + this.cwd)
|
||
error.path = this.cwd
|
||
error.code = er.code
|
||
this.emit('error', error)
|
||
this.abort()
|
||
}
|
||
break
|
||
|
||
case 'ENOENT': // not terribly unusual
|
||
case 'ELOOP':
|
||
case 'ENAMETOOLONG':
|
||
case 'UNKNOWN':
|
||
this.cache[this._makeAbs(f)] = false
|
||
break
|
||
|
||
default: // some unusual error. Treat as failure.
|
||
this.cache[this._makeAbs(f)] = false
|
||
if (this.strict) {
|
||
this.emit('error', er)
|
||
// If the error is handled, then we abort
|
||
// if not, we threw out of here
|
||
this.abort()
|
||
}
|
||
if (!this.silent)
|
||
console.error('glob error', er)
|
||
break
|
||
}
|
||
|
||
return cb()
|
||
}
|
||
|
||
Glob.prototype._processGlobStar = function (prefix, read, abs, remain, index, inGlobStar, cb) {
|
||
var self = this
|
||
this._readdir(abs, inGlobStar, function (er, entries) {
|
||
self._processGlobStar2(prefix, read, abs, remain, index, inGlobStar, entries, cb)
|
||
})
|
||
}
|
||
|
||
|
||
Glob.prototype._processGlobStar2 = function (prefix, read, abs, remain, index, inGlobStar, entries, cb) {
|
||
//console.error('pgs2', prefix, remain[0], entries)
|
||
|
||
// no entries means not a dir, so it can never have matches
|
||
// foo.txt/** doesn't match foo.txt
|
||
if (!entries)
|
||
return cb()
|
||
|
||
// test without the globstar, and with every child both below
|
||
// and replacing the globstar.
|
||
var remainWithoutGlobStar = remain.slice(1)
|
||
var gspref = prefix ? [ prefix ] : []
|
||
var noGlobStar = gspref.concat(remainWithoutGlobStar)
|
||
|
||
// the noGlobStar pattern exits the inGlobStar state
|
||
this._process(noGlobStar, index, false, cb)
|
||
|
||
var isSym = this.symlinks[abs]
|
||
var len = entries.length
|
||
|
||
// If it's a symlink, and we're in a globstar, then stop
|
||
if (isSym && inGlobStar)
|
||
return cb()
|
||
|
||
for (var i = 0; i < len; i++) {
|
||
var e = entries[i]
|
||
if (e.charAt(0) === '.' && !this.dot)
|
||
continue
|
||
|
||
// these two cases enter the inGlobStar state
|
||
var instead = gspref.concat(entries[i], remainWithoutGlobStar)
|
||
this._process(instead, index, true, cb)
|
||
|
||
var below = gspref.concat(entries[i], remain)
|
||
this._process(below, index, true, cb)
|
||
}
|
||
|
||
cb()
|
||
}
|
||
|
||
Glob.prototype._processSimple = function (prefix, index, cb) {
|
||
// XXX review this. Shouldn't it be doing the mounting etc
|
||
// before doing stat? kinda weird?
|
||
var self = this
|
||
this._stat(prefix, function (er, exists) {
|
||
self._processSimple2(prefix, index, er, exists, cb)
|
||
})
|
||
}
|
||
Glob.prototype._processSimple2 = function (prefix, index, er, exists, cb) {
|
||
|
||
//console.error('ps2', prefix, exists)
|
||
|
||
if (!this.matches[index])
|
||
this.matches[index] = Object.create(null)
|
||
|
||
// If it doesn't exist, then just mark the lack of results
|
||
if (!exists)
|
||
return cb()
|
||
|
||
if (prefix && isAbsolute(prefix) && !this.nomount) {
|
||
var trail = /[\/\\]$/.test(prefix)
|
||
if (prefix.charAt(0) === '/') {
|
||
prefix = path.join(this.root, prefix)
|
||
} else {
|
||
prefix = path.resolve(this.root, prefix)
|
||
if (trail)
|
||
prefix += '/'
|
||
}
|
||
}
|
||
|
||
if (process.platform === 'win32')
|
||
prefix = prefix.replace(/\\/g, '/')
|
||
|
||
// Mark this as a match
|
||
this._emitMatch(index, prefix)
|
||
cb()
|
||
}
|
||
|
||
// Returns either 'DIR', 'FILE', or false
|
||
Glob.prototype._stat = function (f, cb) {
|
||
var abs = this._makeAbs(f)
|
||
var needDir = f.slice(-1) === '/'
|
||
|
||
if (f.length > this.maxLength)
|
||
return cb()
|
||
|
||
if (!this.stat && ownProp(this.cache, abs)) {
|
||
var c = this.cache[abs]
|
||
|
||
if (Array.isArray(c))
|
||
c = 'DIR'
|
||
|
||
// It exists, but maybe not how we need it
|
||
if (!needDir || c === 'DIR')
|
||
return cb(null, c)
|
||
|
||
if (needDir && c === 'FILE')
|
||
return cb()
|
||
|
||
// otherwise we have to stat, because maybe c=true
|
||
// if we know it exists, but not what it is.
|
||
}
|
||
|
||
var exists
|
||
var stat = this.statCache[abs]
|
||
if (stat !== undefined) {
|
||
if (stat === false)
|
||
return cb(null, stat)
|
||
else {
|
||
var type = stat.isDirectory() ? 'DIR' : 'FILE'
|
||
if (needDir && type === 'FILE')
|
||
return cb()
|
||
else
|
||
return cb(null, type, stat)
|
||
}
|
||
}
|
||
|
||
var self = this
|
||
var statcb = inflight('stat\0' + abs, lstatcb_)
|
||
if (statcb)
|
||
fs.lstat(abs, statcb)
|
||
|
||
function lstatcb_ (er, lstat) {
|
||
if (lstat && lstat.isSymbolicLink()) {
|
||
// If it's a symlink, then treat it as the target, unless
|
||
// the target does not exist, then treat it as a file.
|
||
return fs.stat(abs, function (er, stat) {
|
||
if (er)
|
||
self._stat2(f, abs, null, lstat, cb)
|
||
else
|
||
self._stat2(f, abs, er, stat, cb)
|
||
})
|
||
} else {
|
||
self._stat2(f, abs, er, lstat, cb)
|
||
}
|
||
}
|
||
}
|
||
|
||
Glob.prototype._stat2 = function (f, abs, er, stat, cb) {
|
||
if (er && (er.code === 'ENOENT' || er.code === 'ENOTDIR')) {
|
||
this.statCache[abs] = false
|
||
return cb()
|
||
}
|
||
|
||
var needDir = f.slice(-1) === '/'
|
||
this.statCache[abs] = stat
|
||
|
||
if (abs.slice(-1) === '/' && stat && !stat.isDirectory())
|
||
return cb(null, false, stat)
|
||
|
||
var c = true
|
||
if (stat)
|
||
c = stat.isDirectory() ? 'DIR' : 'FILE'
|
||
this.cache[abs] = this.cache[abs] || c
|
||
|
||
if (needDir && c === 'FILE')
|
||
return cb()
|
||
|
||
return cb(null, c, stat)
|
||
}
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 411:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var utils = __webpack_require__(35);
|
||
|
||
module.exports = function normalizeHeaderName(headers, normalizedName) {
|
||
utils.forEach(headers, function processHeader(value, name) {
|
||
if (name !== normalizedName && name.toUpperCase() === normalizedName.toUpperCase()) {
|
||
headers[normalizedName] = value;
|
||
delete headers[name];
|
||
}
|
||
});
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 413:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("stream");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 417:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("crypto");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 431:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __importStar = (this && this.__importStar) || function (mod) {
|
||
if (mod && mod.__esModule) return mod;
|
||
var result = {};
|
||
if (mod != null) for (var k in mod) if (Object.hasOwnProperty.call(mod, k)) result[k] = mod[k];
|
||
result["default"] = mod;
|
||
return result;
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const os = __importStar(__webpack_require__(87));
|
||
const utils_1 = __webpack_require__(82);
|
||
/**
|
||
* Commands
|
||
*
|
||
* Command Format:
|
||
* ::name key=value,key=value::message
|
||
*
|
||
* Examples:
|
||
* ::warning::This is the message
|
||
* ::set-env name=MY_VAR::some value
|
||
*/
|
||
function issueCommand(command, properties, message) {
|
||
const cmd = new Command(command, properties, message);
|
||
process.stdout.write(cmd.toString() + os.EOL);
|
||
}
|
||
exports.issueCommand = issueCommand;
|
||
function issue(name, message = '') {
|
||
issueCommand(name, {}, message);
|
||
}
|
||
exports.issue = issue;
|
||
const CMD_STRING = '::';
|
||
class Command {
|
||
constructor(command, properties, message) {
|
||
if (!command) {
|
||
command = 'missing.command';
|
||
}
|
||
this.command = command;
|
||
this.properties = properties;
|
||
this.message = message;
|
||
}
|
||
toString() {
|
||
let cmdStr = CMD_STRING + this.command;
|
||
if (this.properties && Object.keys(this.properties).length > 0) {
|
||
cmdStr += ' ';
|
||
let first = true;
|
||
for (const key in this.properties) {
|
||
if (this.properties.hasOwnProperty(key)) {
|
||
const val = this.properties[key];
|
||
if (val) {
|
||
if (first) {
|
||
first = false;
|
||
}
|
||
else {
|
||
cmdStr += ',';
|
||
}
|
||
cmdStr += `${key}=${escapeProperty(val)}`;
|
||
}
|
||
}
|
||
}
|
||
}
|
||
cmdStr += `${CMD_STRING}${escapeData(this.message)}`;
|
||
return cmdStr;
|
||
}
|
||
}
|
||
function escapeData(s) {
|
||
return utils_1.toCommandValue(s)
|
||
.replace(/%/g, '%25')
|
||
.replace(/\r/g, '%0D')
|
||
.replace(/\n/g, '%0A');
|
||
}
|
||
function escapeProperty(s) {
|
||
return utils_1.toCommandValue(s)
|
||
.replace(/%/g, '%25')
|
||
.replace(/\r/g, '%0D')
|
||
.replace(/\n/g, '%0A')
|
||
.replace(/:/g, '%3A')
|
||
.replace(/,/g, '%2C');
|
||
}
|
||
//# sourceMappingURL=command.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 452:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const utils_1 = __webpack_require__(870);
|
||
/**
|
||
* Used for managing http clients during either upload or download
|
||
*/
|
||
class HttpManager {
|
||
constructor(clientCount, userAgent) {
|
||
if (clientCount < 1) {
|
||
throw new Error('There must be at least one client');
|
||
}
|
||
this.userAgent = userAgent;
|
||
this.clients = new Array(clientCount).fill(utils_1.createHttpClient(userAgent));
|
||
}
|
||
getClient(index) {
|
||
return this.clients[index];
|
||
}
|
||
// client disposal is necessary if a keep-alive connection is used to properly close the connection
|
||
// for more information see: https://github.com/actions/http-client/blob/04e5ad73cd3fd1f5610a32116b0759eddf6570d2/index.ts#L292
|
||
disposeAndReplaceClient(index) {
|
||
this.clients[index].dispose();
|
||
this.clients[index] = utils_1.createHttpClient(this.userAgent);
|
||
}
|
||
disposeAndReplaceAllClients() {
|
||
for (const [index] of this.clients.entries()) {
|
||
this.disposeAndReplaceClient(index);
|
||
}
|
||
}
|
||
}
|
||
exports.HttpManager = HttpManager;
|
||
//# sourceMappingURL=http-manager.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 454:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
var debug;
|
||
|
||
module.exports = function () {
|
||
if (!debug) {
|
||
try {
|
||
/* eslint global-require: off */
|
||
debug = __webpack_require__(784)("follow-redirects");
|
||
}
|
||
catch (error) { /* */ }
|
||
if (typeof debug !== "function") {
|
||
debug = function () { /* */ };
|
||
}
|
||
}
|
||
debug.apply(null, arguments);
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 470:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
||
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
||
return new (P || (P = Promise))(function (resolve, reject) {
|
||
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
||
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
||
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
||
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
||
});
|
||
};
|
||
var __importStar = (this && this.__importStar) || function (mod) {
|
||
if (mod && mod.__esModule) return mod;
|
||
var result = {};
|
||
if (mod != null) for (var k in mod) if (Object.hasOwnProperty.call(mod, k)) result[k] = mod[k];
|
||
result["default"] = mod;
|
||
return result;
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const command_1 = __webpack_require__(431);
|
||
const file_command_1 = __webpack_require__(102);
|
||
const utils_1 = __webpack_require__(82);
|
||
const os = __importStar(__webpack_require__(87));
|
||
const path = __importStar(__webpack_require__(622));
|
||
/**
|
||
* The code to exit an action
|
||
*/
|
||
var ExitCode;
|
||
(function (ExitCode) {
|
||
/**
|
||
* A code indicating that the action was successful
|
||
*/
|
||
ExitCode[ExitCode["Success"] = 0] = "Success";
|
||
/**
|
||
* A code indicating that the action was a failure
|
||
*/
|
||
ExitCode[ExitCode["Failure"] = 1] = "Failure";
|
||
})(ExitCode = exports.ExitCode || (exports.ExitCode = {}));
|
||
//-----------------------------------------------------------------------
|
||
// Variables
|
||
//-----------------------------------------------------------------------
|
||
/**
|
||
* Sets env variable for this action and future actions in the job
|
||
* @param name the name of the variable to set
|
||
* @param val the value of the variable. Non-string values will be converted to a string via JSON.stringify
|
||
*/
|
||
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
||
function exportVariable(name, val) {
|
||
const convertedVal = utils_1.toCommandValue(val);
|
||
process.env[name] = convertedVal;
|
||
const filePath = process.env['GITHUB_ENV'] || '';
|
||
if (filePath) {
|
||
const delimiter = '_GitHubActionsFileCommandDelimeter_';
|
||
const commandValue = `${name}<<${delimiter}${os.EOL}${convertedVal}${os.EOL}${delimiter}`;
|
||
file_command_1.issueCommand('ENV', commandValue);
|
||
}
|
||
else {
|
||
command_1.issueCommand('set-env', { name }, convertedVal);
|
||
}
|
||
}
|
||
exports.exportVariable = exportVariable;
|
||
/**
|
||
* Registers a secret which will get masked from logs
|
||
* @param secret value of the secret
|
||
*/
|
||
function setSecret(secret) {
|
||
command_1.issueCommand('add-mask', {}, secret);
|
||
}
|
||
exports.setSecret = setSecret;
|
||
/**
|
||
* Prepends inputPath to the PATH (for this action and future actions)
|
||
* @param inputPath
|
||
*/
|
||
function addPath(inputPath) {
|
||
const filePath = process.env['GITHUB_PATH'] || '';
|
||
if (filePath) {
|
||
file_command_1.issueCommand('PATH', inputPath);
|
||
}
|
||
else {
|
||
command_1.issueCommand('add-path', {}, inputPath);
|
||
}
|
||
process.env['PATH'] = `${inputPath}${path.delimiter}${process.env['PATH']}`;
|
||
}
|
||
exports.addPath = addPath;
|
||
/**
|
||
* Gets the value of an input. The value is also trimmed.
|
||
*
|
||
* @param name name of the input to get
|
||
* @param options optional. See InputOptions.
|
||
* @returns string
|
||
*/
|
||
function getInput(name, options) {
|
||
const val = process.env[`INPUT_${name.replace(/ /g, '_').toUpperCase()}`] || '';
|
||
if (options && options.required && !val) {
|
||
throw new Error(`Input required and not supplied: ${name}`);
|
||
}
|
||
return val.trim();
|
||
}
|
||
exports.getInput = getInput;
|
||
/**
|
||
* Sets the value of an output.
|
||
*
|
||
* @param name name of the output to set
|
||
* @param value value to store. Non-string values will be converted to a string via JSON.stringify
|
||
*/
|
||
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
||
function setOutput(name, value) {
|
||
command_1.issueCommand('set-output', { name }, value);
|
||
}
|
||
exports.setOutput = setOutput;
|
||
/**
|
||
* Enables or disables the echoing of commands into stdout for the rest of the step.
|
||
* Echoing is disabled by default if ACTIONS_STEP_DEBUG is not set.
|
||
*
|
||
*/
|
||
function setCommandEcho(enabled) {
|
||
command_1.issue('echo', enabled ? 'on' : 'off');
|
||
}
|
||
exports.setCommandEcho = setCommandEcho;
|
||
//-----------------------------------------------------------------------
|
||
// Results
|
||
//-----------------------------------------------------------------------
|
||
/**
|
||
* Sets the action status to failed.
|
||
* When the action exits it will be with an exit code of 1
|
||
* @param message add error issue message
|
||
*/
|
||
function setFailed(message) {
|
||
process.exitCode = ExitCode.Failure;
|
||
error(message);
|
||
}
|
||
exports.setFailed = setFailed;
|
||
//-----------------------------------------------------------------------
|
||
// Logging Commands
|
||
//-----------------------------------------------------------------------
|
||
/**
|
||
* Gets whether Actions Step Debug is on or not
|
||
*/
|
||
function isDebug() {
|
||
return process.env['RUNNER_DEBUG'] === '1';
|
||
}
|
||
exports.isDebug = isDebug;
|
||
/**
|
||
* Writes debug message to user log
|
||
* @param message debug message
|
||
*/
|
||
function debug(message) {
|
||
command_1.issueCommand('debug', {}, message);
|
||
}
|
||
exports.debug = debug;
|
||
/**
|
||
* Adds an error issue
|
||
* @param message error issue message. Errors will be converted to string via toString()
|
||
*/
|
||
function error(message) {
|
||
command_1.issue('error', message instanceof Error ? message.toString() : message);
|
||
}
|
||
exports.error = error;
|
||
/**
|
||
* Adds an warning issue
|
||
* @param message warning issue message. Errors will be converted to string via toString()
|
||
*/
|
||
function warning(message) {
|
||
command_1.issue('warning', message instanceof Error ? message.toString() : message);
|
||
}
|
||
exports.warning = warning;
|
||
/**
|
||
* Writes info to log with console.log.
|
||
* @param message info message
|
||
*/
|
||
function info(message) {
|
||
process.stdout.write(message + os.EOL);
|
||
}
|
||
exports.info = info;
|
||
/**
|
||
* Begin an output group.
|
||
*
|
||
* Output until the next `groupEnd` will be foldable in this group
|
||
*
|
||
* @param name The name of the output group
|
||
*/
|
||
function startGroup(name) {
|
||
command_1.issue('group', name);
|
||
}
|
||
exports.startGroup = startGroup;
|
||
/**
|
||
* End an output group.
|
||
*/
|
||
function endGroup() {
|
||
command_1.issue('endgroup');
|
||
}
|
||
exports.endGroup = endGroup;
|
||
/**
|
||
* Wrap an asynchronous function call in a group.
|
||
*
|
||
* Returns the same type as the function itself.
|
||
*
|
||
* @param name The name of the group
|
||
* @param fn The function to wrap in the group
|
||
*/
|
||
function group(name, fn) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
startGroup(name);
|
||
let result;
|
||
try {
|
||
result = yield fn();
|
||
}
|
||
finally {
|
||
endGroup();
|
||
}
|
||
return result;
|
||
});
|
||
}
|
||
exports.group = group;
|
||
//-----------------------------------------------------------------------
|
||
// Wrapper action state
|
||
//-----------------------------------------------------------------------
|
||
/**
|
||
* Saves state for current action, the state can only be retrieved by this action's post job execution.
|
||
*
|
||
* @param name name of the state to store
|
||
* @param value value to store. Non-string values will be converted to a string via JSON.stringify
|
||
*/
|
||
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
||
function saveState(name, value) {
|
||
command_1.issueCommand('save-state', { name }, value);
|
||
}
|
||
exports.saveState = saveState;
|
||
/**
|
||
* Gets the value of an state set by this action's main execution.
|
||
*
|
||
* @param name name of the state to get
|
||
* @returns string
|
||
*/
|
||
function getState(name) {
|
||
return process.env[`STATE_${name}`] || '';
|
||
}
|
||
exports.getState = getState;
|
||
//# sourceMappingURL=core.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 486:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
|
||
/**
|
||
* This is the common logic for both the Node.js and web browser
|
||
* implementations of `debug()`.
|
||
*/
|
||
|
||
function setup(env) {
|
||
createDebug.debug = createDebug;
|
||
createDebug.default = createDebug;
|
||
createDebug.coerce = coerce;
|
||
createDebug.disable = disable;
|
||
createDebug.enable = enable;
|
||
createDebug.enabled = enabled;
|
||
createDebug.humanize = __webpack_require__(761);
|
||
|
||
Object.keys(env).forEach(key => {
|
||
createDebug[key] = env[key];
|
||
});
|
||
|
||
/**
|
||
* Active `debug` instances.
|
||
*/
|
||
createDebug.instances = [];
|
||
|
||
/**
|
||
* The currently active debug mode names, and names to skip.
|
||
*/
|
||
|
||
createDebug.names = [];
|
||
createDebug.skips = [];
|
||
|
||
/**
|
||
* Map of special "%n" handling functions, for the debug "format" argument.
|
||
*
|
||
* Valid key names are a single, lower or upper-case letter, i.e. "n" and "N".
|
||
*/
|
||
createDebug.formatters = {};
|
||
|
||
/**
|
||
* Selects a color for a debug namespace
|
||
* @param {String} namespace The namespace string for the for the debug instance to be colored
|
||
* @return {Number|String} An ANSI color code for the given namespace
|
||
* @api private
|
||
*/
|
||
function selectColor(namespace) {
|
||
let hash = 0;
|
||
|
||
for (let i = 0; i < namespace.length; i++) {
|
||
hash = ((hash << 5) - hash) + namespace.charCodeAt(i);
|
||
hash |= 0; // Convert to 32bit integer
|
||
}
|
||
|
||
return createDebug.colors[Math.abs(hash) % createDebug.colors.length];
|
||
}
|
||
createDebug.selectColor = selectColor;
|
||
|
||
/**
|
||
* Create a debugger with the given `namespace`.
|
||
*
|
||
* @param {String} namespace
|
||
* @return {Function}
|
||
* @api public
|
||
*/
|
||
function createDebug(namespace) {
|
||
let prevTime;
|
||
|
||
function debug(...args) {
|
||
// Disabled?
|
||
if (!debug.enabled) {
|
||
return;
|
||
}
|
||
|
||
const self = debug;
|
||
|
||
// Set `diff` timestamp
|
||
const curr = Number(new Date());
|
||
const ms = curr - (prevTime || curr);
|
||
self.diff = ms;
|
||
self.prev = prevTime;
|
||
self.curr = curr;
|
||
prevTime = curr;
|
||
|
||
args[0] = createDebug.coerce(args[0]);
|
||
|
||
if (typeof args[0] !== 'string') {
|
||
// Anything else let's inspect with %O
|
||
args.unshift('%O');
|
||
}
|
||
|
||
// Apply any `formatters` transformations
|
||
let index = 0;
|
||
args[0] = args[0].replace(/%([a-zA-Z%])/g, (match, format) => {
|
||
// If we encounter an escaped % then don't increase the array index
|
||
if (match === '%%') {
|
||
return match;
|
||
}
|
||
index++;
|
||
const formatter = createDebug.formatters[format];
|
||
if (typeof formatter === 'function') {
|
||
const val = args[index];
|
||
match = formatter.call(self, val);
|
||
|
||
// Now we need to remove `args[index]` since it's inlined in the `format`
|
||
args.splice(index, 1);
|
||
index--;
|
||
}
|
||
return match;
|
||
});
|
||
|
||
// Apply env-specific formatting (colors, etc.)
|
||
createDebug.formatArgs.call(self, args);
|
||
|
||
const logFn = self.log || createDebug.log;
|
||
logFn.apply(self, args);
|
||
}
|
||
|
||
debug.namespace = namespace;
|
||
debug.enabled = createDebug.enabled(namespace);
|
||
debug.useColors = createDebug.useColors();
|
||
debug.color = selectColor(namespace);
|
||
debug.destroy = destroy;
|
||
debug.extend = extend;
|
||
// Debug.formatArgs = formatArgs;
|
||
// debug.rawLog = rawLog;
|
||
|
||
// env-specific initialization logic for debug instances
|
||
if (typeof createDebug.init === 'function') {
|
||
createDebug.init(debug);
|
||
}
|
||
|
||
createDebug.instances.push(debug);
|
||
|
||
return debug;
|
||
}
|
||
|
||
function destroy() {
|
||
const index = createDebug.instances.indexOf(this);
|
||
if (index !== -1) {
|
||
createDebug.instances.splice(index, 1);
|
||
return true;
|
||
}
|
||
return false;
|
||
}
|
||
|
||
function extend(namespace, delimiter) {
|
||
const newDebug = createDebug(this.namespace + (typeof delimiter === 'undefined' ? ':' : delimiter) + namespace);
|
||
newDebug.log = this.log;
|
||
return newDebug;
|
||
}
|
||
|
||
/**
|
||
* Enables a debug mode by namespaces. This can include modes
|
||
* separated by a colon and wildcards.
|
||
*
|
||
* @param {String} namespaces
|
||
* @api public
|
||
*/
|
||
function enable(namespaces) {
|
||
createDebug.save(namespaces);
|
||
|
||
createDebug.names = [];
|
||
createDebug.skips = [];
|
||
|
||
let i;
|
||
const split = (typeof namespaces === 'string' ? namespaces : '').split(/[\s,]+/);
|
||
const len = split.length;
|
||
|
||
for (i = 0; i < len; i++) {
|
||
if (!split[i]) {
|
||
// ignore empty strings
|
||
continue;
|
||
}
|
||
|
||
namespaces = split[i].replace(/\*/g, '.*?');
|
||
|
||
if (namespaces[0] === '-') {
|
||
createDebug.skips.push(new RegExp('^' + namespaces.substr(1) + '$'));
|
||
} else {
|
||
createDebug.names.push(new RegExp('^' + namespaces + '$'));
|
||
}
|
||
}
|
||
|
||
for (i = 0; i < createDebug.instances.length; i++) {
|
||
const instance = createDebug.instances[i];
|
||
instance.enabled = createDebug.enabled(instance.namespace);
|
||
}
|
||
}
|
||
|
||
/**
|
||
* Disable debug output.
|
||
*
|
||
* @return {String} namespaces
|
||
* @api public
|
||
*/
|
||
function disable() {
|
||
const namespaces = [
|
||
...createDebug.names.map(toNamespace),
|
||
...createDebug.skips.map(toNamespace).map(namespace => '-' + namespace)
|
||
].join(',');
|
||
createDebug.enable('');
|
||
return namespaces;
|
||
}
|
||
|
||
/**
|
||
* Returns true if the given mode name is enabled, false otherwise.
|
||
*
|
||
* @param {String} name
|
||
* @return {Boolean}
|
||
* @api public
|
||
*/
|
||
function enabled(name) {
|
||
if (name[name.length - 1] === '*') {
|
||
return true;
|
||
}
|
||
|
||
let i;
|
||
let len;
|
||
|
||
for (i = 0, len = createDebug.skips.length; i < len; i++) {
|
||
if (createDebug.skips[i].test(name)) {
|
||
return false;
|
||
}
|
||
}
|
||
|
||
for (i = 0, len = createDebug.names.length; i < len; i++) {
|
||
if (createDebug.names[i].test(name)) {
|
||
return true;
|
||
}
|
||
}
|
||
|
||
return false;
|
||
}
|
||
|
||
/**
|
||
* Convert regexp to namespace
|
||
*
|
||
* @param {RegExp} regxep
|
||
* @return {String} namespace
|
||
* @api private
|
||
*/
|
||
function toNamespace(regexp) {
|
||
return regexp.toString()
|
||
.substring(2, regexp.toString().length - 2)
|
||
.replace(/\.\*\?$/, '*');
|
||
}
|
||
|
||
/**
|
||
* Coerce `val`.
|
||
*
|
||
* @param {Mixed} val
|
||
* @return {Mixed}
|
||
* @api private
|
||
*/
|
||
function coerce(val) {
|
||
if (val instanceof Error) {
|
||
return val.stack || val.message;
|
||
}
|
||
return val;
|
||
}
|
||
|
||
createDebug.enable(createDebug.load());
|
||
|
||
return createDebug;
|
||
}
|
||
|
||
module.exports = setup;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 489:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
||
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
||
return new (P || (P = Promise))(function (resolve, reject) {
|
||
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
||
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
||
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
||
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
||
});
|
||
};
|
||
var __importStar = (this && this.__importStar) || function (mod) {
|
||
if (mod && mod.__esModule) return mod;
|
||
var result = {};
|
||
if (mod != null) for (var k in mod) if (Object.hasOwnProperty.call(mod, k)) result[k] = mod[k];
|
||
result["default"] = mod;
|
||
return result;
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const utils_1 = __webpack_require__(870);
|
||
const core = __importStar(__webpack_require__(470));
|
||
const config_variables_1 = __webpack_require__(401);
|
||
function retry(name, operation, customErrorMessages, maxAttempts) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
let response = undefined;
|
||
let statusCode = undefined;
|
||
let isRetryable = false;
|
||
let errorMessage = '';
|
||
let customErrorInformation = undefined;
|
||
let attempt = 1;
|
||
while (attempt <= maxAttempts) {
|
||
try {
|
||
response = yield operation();
|
||
statusCode = response.message.statusCode;
|
||
if (utils_1.isSuccessStatusCode(statusCode)) {
|
||
return response;
|
||
}
|
||
// Extra error information that we want to display if a particular response code is hit
|
||
if (statusCode) {
|
||
customErrorInformation = customErrorMessages.get(statusCode);
|
||
}
|
||
isRetryable = utils_1.isRetryableStatusCode(statusCode);
|
||
errorMessage = `Artifact service responded with ${statusCode}`;
|
||
}
|
||
catch (error) {
|
||
isRetryable = true;
|
||
errorMessage = error.message;
|
||
}
|
||
if (!isRetryable) {
|
||
core.info(`${name} - Error is not retryable`);
|
||
if (response) {
|
||
utils_1.displayHttpDiagnostics(response);
|
||
}
|
||
break;
|
||
}
|
||
core.info(`${name} - Attempt ${attempt} of ${maxAttempts} failed with error: ${errorMessage}`);
|
||
yield utils_1.sleep(utils_1.getExponentialRetryTimeInMilliseconds(attempt));
|
||
attempt++;
|
||
}
|
||
if (response) {
|
||
utils_1.displayHttpDiagnostics(response);
|
||
}
|
||
if (customErrorInformation) {
|
||
throw Error(`${name} failed: ${customErrorInformation}`);
|
||
}
|
||
throw Error(`${name} failed: ${errorMessage}`);
|
||
});
|
||
}
|
||
exports.retry = retry;
|
||
function retryHttpClientRequest(name, method, customErrorMessages = new Map(), maxAttempts = config_variables_1.getRetryLimit()) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
return yield retry(name, method, customErrorMessages, maxAttempts);
|
||
});
|
||
}
|
||
exports.retryHttpClientRequest = retryHttpClientRequest;
|
||
//# sourceMappingURL=requestUtils.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 529:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var utils = __webpack_require__(35);
|
||
var normalizeHeaderName = __webpack_require__(411);
|
||
var enhanceError = __webpack_require__(369);
|
||
|
||
var DEFAULT_CONTENT_TYPE = {
|
||
'Content-Type': 'application/x-www-form-urlencoded'
|
||
};
|
||
|
||
function setContentTypeIfUnset(headers, value) {
|
||
if (!utils.isUndefined(headers) && utils.isUndefined(headers['Content-Type'])) {
|
||
headers['Content-Type'] = value;
|
||
}
|
||
}
|
||
|
||
function getDefaultAdapter() {
|
||
var adapter;
|
||
if (typeof XMLHttpRequest !== 'undefined') {
|
||
// For browsers use XHR adapter
|
||
adapter = __webpack_require__(219);
|
||
} else if (typeof process !== 'undefined' && Object.prototype.toString.call(process) === '[object process]') {
|
||
// For node use HTTP adapter
|
||
adapter = __webpack_require__(670);
|
||
}
|
||
return adapter;
|
||
}
|
||
|
||
function stringifySafely(rawValue, parser, encoder) {
|
||
if (utils.isString(rawValue)) {
|
||
try {
|
||
(parser || JSON.parse)(rawValue);
|
||
return utils.trim(rawValue);
|
||
} catch (e) {
|
||
if (e.name !== 'SyntaxError') {
|
||
throw e;
|
||
}
|
||
}
|
||
}
|
||
|
||
return (encoder || JSON.stringify)(rawValue);
|
||
}
|
||
|
||
var defaults = {
|
||
|
||
transitional: {
|
||
silentJSONParsing: true,
|
||
forcedJSONParsing: true,
|
||
clarifyTimeoutError: false
|
||
},
|
||
|
||
adapter: getDefaultAdapter(),
|
||
|
||
transformRequest: [function transformRequest(data, headers) {
|
||
normalizeHeaderName(headers, 'Accept');
|
||
normalizeHeaderName(headers, 'Content-Type');
|
||
|
||
if (utils.isFormData(data) ||
|
||
utils.isArrayBuffer(data) ||
|
||
utils.isBuffer(data) ||
|
||
utils.isStream(data) ||
|
||
utils.isFile(data) ||
|
||
utils.isBlob(data)
|
||
) {
|
||
return data;
|
||
}
|
||
if (utils.isArrayBufferView(data)) {
|
||
return data.buffer;
|
||
}
|
||
if (utils.isURLSearchParams(data)) {
|
||
setContentTypeIfUnset(headers, 'application/x-www-form-urlencoded;charset=utf-8');
|
||
return data.toString();
|
||
}
|
||
if (utils.isObject(data) || (headers && headers['Content-Type'] === 'application/json')) {
|
||
setContentTypeIfUnset(headers, 'application/json');
|
||
return stringifySafely(data);
|
||
}
|
||
return data;
|
||
}],
|
||
|
||
transformResponse: [function transformResponse(data) {
|
||
var transitional = this.transitional;
|
||
var silentJSONParsing = transitional && transitional.silentJSONParsing;
|
||
var forcedJSONParsing = transitional && transitional.forcedJSONParsing;
|
||
var strictJSONParsing = !silentJSONParsing && this.responseType === 'json';
|
||
|
||
if (strictJSONParsing || (forcedJSONParsing && utils.isString(data) && data.length)) {
|
||
try {
|
||
return JSON.parse(data);
|
||
} catch (e) {
|
||
if (strictJSONParsing) {
|
||
if (e.name === 'SyntaxError') {
|
||
throw enhanceError(e, this, 'E_JSON_PARSE');
|
||
}
|
||
throw e;
|
||
}
|
||
}
|
||
}
|
||
|
||
return data;
|
||
}],
|
||
|
||
/**
|
||
* A timeout in milliseconds to abort a request. If set to 0 (default) a
|
||
* timeout is not created.
|
||
*/
|
||
timeout: 0,
|
||
|
||
xsrfCookieName: 'XSRF-TOKEN',
|
||
xsrfHeaderName: 'X-XSRF-TOKEN',
|
||
|
||
maxContentLength: -1,
|
||
maxBodyLength: -1,
|
||
|
||
validateStatus: function validateStatus(status) {
|
||
return status >= 200 && status < 300;
|
||
}
|
||
};
|
||
|
||
defaults.headers = {
|
||
common: {
|
||
'Accept': 'application/json, text/plain, */*'
|
||
}
|
||
};
|
||
|
||
utils.forEach(['delete', 'get', 'head'], function forEachMethodNoData(method) {
|
||
defaults.headers[method] = {};
|
||
});
|
||
|
||
utils.forEach(['post', 'put', 'patch'], function forEachMethodWithData(method) {
|
||
defaults.headers[method] = utils.merge(DEFAULT_CONTENT_TYPE);
|
||
});
|
||
|
||
module.exports = defaults;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 532:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __importStar = (this && this.__importStar) || function (mod) {
|
||
if (mod && mod.__esModule) return mod;
|
||
var result = {};
|
||
if (mod != null) for (var k in mod) if (Object.hasOwnProperty.call(mod, k)) result[k] = mod[k];
|
||
result["default"] = mod;
|
||
return result;
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const path = __importStar(__webpack_require__(622));
|
||
/**
|
||
* Creates a specification for a set of files that will be downloaded
|
||
* @param artifactName the name of the artifact
|
||
* @param artifactEntries a set of container entries that describe that files that make up an artifact
|
||
* @param downloadPath the path where the artifact will be downloaded to
|
||
* @param includeRootDirectory specifies if there should be an extra directory (denoted by the artifact name) where the artifact files should be downloaded to
|
||
*/
|
||
function getDownloadSpecification(artifactName, artifactEntries, downloadPath, includeRootDirectory) {
|
||
// use a set for the directory paths so that there are no duplicates
|
||
const directories = new Set();
|
||
const specifications = {
|
||
rootDownloadLocation: includeRootDirectory
|
||
? path.join(downloadPath, artifactName)
|
||
: downloadPath,
|
||
directoryStructure: [],
|
||
emptyFilesToCreate: [],
|
||
filesToDownload: []
|
||
};
|
||
for (const entry of artifactEntries) {
|
||
// Ignore artifacts in the container that don't begin with the same name
|
||
if (entry.path.startsWith(`${artifactName}/`) ||
|
||
entry.path.startsWith(`${artifactName}\\`)) {
|
||
// normalize all separators to the local OS
|
||
const normalizedPathEntry = path.normalize(entry.path);
|
||
// entry.path always starts with the artifact name, if includeRootDirectory is false, remove the name from the beginning of the path
|
||
const filePath = path.join(downloadPath, includeRootDirectory
|
||
? normalizedPathEntry
|
||
: normalizedPathEntry.replace(artifactName, ''));
|
||
// Case insensitive folder structure maintained in the backend, not every folder is created so the 'folder'
|
||
// itemType cannot be relied upon. The file must be used to determine the directory structure
|
||
if (entry.itemType === 'file') {
|
||
// Get the directories that we need to create from the filePath for each individual file
|
||
directories.add(path.dirname(filePath));
|
||
if (entry.fileLength === 0) {
|
||
// An empty file was uploaded, create the empty files locally so that no extra http calls are made
|
||
specifications.emptyFilesToCreate.push(filePath);
|
||
}
|
||
else {
|
||
specifications.filesToDownload.push({
|
||
sourceLocation: entry.contentLocation,
|
||
targetPath: filePath
|
||
});
|
||
}
|
||
}
|
||
}
|
||
}
|
||
specifications.directoryStructure = Array.from(directories);
|
||
return specifications;
|
||
}
|
||
exports.getDownloadSpecification = getDownloadSpecification;
|
||
//# sourceMappingURL=download-specification.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 539:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const http = __webpack_require__(605);
|
||
const https = __webpack_require__(211);
|
||
const pm = __webpack_require__(950);
|
||
let tunnel;
|
||
var HttpCodes;
|
||
(function (HttpCodes) {
|
||
HttpCodes[HttpCodes["OK"] = 200] = "OK";
|
||
HttpCodes[HttpCodes["MultipleChoices"] = 300] = "MultipleChoices";
|
||
HttpCodes[HttpCodes["MovedPermanently"] = 301] = "MovedPermanently";
|
||
HttpCodes[HttpCodes["ResourceMoved"] = 302] = "ResourceMoved";
|
||
HttpCodes[HttpCodes["SeeOther"] = 303] = "SeeOther";
|
||
HttpCodes[HttpCodes["NotModified"] = 304] = "NotModified";
|
||
HttpCodes[HttpCodes["UseProxy"] = 305] = "UseProxy";
|
||
HttpCodes[HttpCodes["SwitchProxy"] = 306] = "SwitchProxy";
|
||
HttpCodes[HttpCodes["TemporaryRedirect"] = 307] = "TemporaryRedirect";
|
||
HttpCodes[HttpCodes["PermanentRedirect"] = 308] = "PermanentRedirect";
|
||
HttpCodes[HttpCodes["BadRequest"] = 400] = "BadRequest";
|
||
HttpCodes[HttpCodes["Unauthorized"] = 401] = "Unauthorized";
|
||
HttpCodes[HttpCodes["PaymentRequired"] = 402] = "PaymentRequired";
|
||
HttpCodes[HttpCodes["Forbidden"] = 403] = "Forbidden";
|
||
HttpCodes[HttpCodes["NotFound"] = 404] = "NotFound";
|
||
HttpCodes[HttpCodes["MethodNotAllowed"] = 405] = "MethodNotAllowed";
|
||
HttpCodes[HttpCodes["NotAcceptable"] = 406] = "NotAcceptable";
|
||
HttpCodes[HttpCodes["ProxyAuthenticationRequired"] = 407] = "ProxyAuthenticationRequired";
|
||
HttpCodes[HttpCodes["RequestTimeout"] = 408] = "RequestTimeout";
|
||
HttpCodes[HttpCodes["Conflict"] = 409] = "Conflict";
|
||
HttpCodes[HttpCodes["Gone"] = 410] = "Gone";
|
||
HttpCodes[HttpCodes["TooManyRequests"] = 429] = "TooManyRequests";
|
||
HttpCodes[HttpCodes["InternalServerError"] = 500] = "InternalServerError";
|
||
HttpCodes[HttpCodes["NotImplemented"] = 501] = "NotImplemented";
|
||
HttpCodes[HttpCodes["BadGateway"] = 502] = "BadGateway";
|
||
HttpCodes[HttpCodes["ServiceUnavailable"] = 503] = "ServiceUnavailable";
|
||
HttpCodes[HttpCodes["GatewayTimeout"] = 504] = "GatewayTimeout";
|
||
})(HttpCodes = exports.HttpCodes || (exports.HttpCodes = {}));
|
||
var Headers;
|
||
(function (Headers) {
|
||
Headers["Accept"] = "accept";
|
||
Headers["ContentType"] = "content-type";
|
||
})(Headers = exports.Headers || (exports.Headers = {}));
|
||
var MediaTypes;
|
||
(function (MediaTypes) {
|
||
MediaTypes["ApplicationJson"] = "application/json";
|
||
})(MediaTypes = exports.MediaTypes || (exports.MediaTypes = {}));
|
||
/**
|
||
* Returns the proxy URL, depending upon the supplied url and proxy environment variables.
|
||
* @param serverUrl The server URL where the request will be sent. For example, https://api.github.com
|
||
*/
|
||
function getProxyUrl(serverUrl) {
|
||
let proxyUrl = pm.getProxyUrl(new URL(serverUrl));
|
||
return proxyUrl ? proxyUrl.href : '';
|
||
}
|
||
exports.getProxyUrl = getProxyUrl;
|
||
const HttpRedirectCodes = [
|
||
HttpCodes.MovedPermanently,
|
||
HttpCodes.ResourceMoved,
|
||
HttpCodes.SeeOther,
|
||
HttpCodes.TemporaryRedirect,
|
||
HttpCodes.PermanentRedirect
|
||
];
|
||
const HttpResponseRetryCodes = [
|
||
HttpCodes.BadGateway,
|
||
HttpCodes.ServiceUnavailable,
|
||
HttpCodes.GatewayTimeout
|
||
];
|
||
const RetryableHttpVerbs = ['OPTIONS', 'GET', 'DELETE', 'HEAD'];
|
||
const ExponentialBackoffCeiling = 10;
|
||
const ExponentialBackoffTimeSlice = 5;
|
||
class HttpClientError extends Error {
|
||
constructor(message, statusCode) {
|
||
super(message);
|
||
this.name = 'HttpClientError';
|
||
this.statusCode = statusCode;
|
||
Object.setPrototypeOf(this, HttpClientError.prototype);
|
||
}
|
||
}
|
||
exports.HttpClientError = HttpClientError;
|
||
class HttpClientResponse {
|
||
constructor(message) {
|
||
this.message = message;
|
||
}
|
||
readBody() {
|
||
return new Promise(async (resolve, reject) => {
|
||
let output = Buffer.alloc(0);
|
||
this.message.on('data', (chunk) => {
|
||
output = Buffer.concat([output, chunk]);
|
||
});
|
||
this.message.on('end', () => {
|
||
resolve(output.toString());
|
||
});
|
||
});
|
||
}
|
||
}
|
||
exports.HttpClientResponse = HttpClientResponse;
|
||
function isHttps(requestUrl) {
|
||
let parsedUrl = new URL(requestUrl);
|
||
return parsedUrl.protocol === 'https:';
|
||
}
|
||
exports.isHttps = isHttps;
|
||
class HttpClient {
|
||
constructor(userAgent, handlers, requestOptions) {
|
||
this._ignoreSslError = false;
|
||
this._allowRedirects = true;
|
||
this._allowRedirectDowngrade = false;
|
||
this._maxRedirects = 50;
|
||
this._allowRetries = false;
|
||
this._maxRetries = 1;
|
||
this._keepAlive = false;
|
||
this._disposed = false;
|
||
this.userAgent = userAgent;
|
||
this.handlers = handlers || [];
|
||
this.requestOptions = requestOptions;
|
||
if (requestOptions) {
|
||
if (requestOptions.ignoreSslError != null) {
|
||
this._ignoreSslError = requestOptions.ignoreSslError;
|
||
}
|
||
this._socketTimeout = requestOptions.socketTimeout;
|
||
if (requestOptions.allowRedirects != null) {
|
||
this._allowRedirects = requestOptions.allowRedirects;
|
||
}
|
||
if (requestOptions.allowRedirectDowngrade != null) {
|
||
this._allowRedirectDowngrade = requestOptions.allowRedirectDowngrade;
|
||
}
|
||
if (requestOptions.maxRedirects != null) {
|
||
this._maxRedirects = Math.max(requestOptions.maxRedirects, 0);
|
||
}
|
||
if (requestOptions.keepAlive != null) {
|
||
this._keepAlive = requestOptions.keepAlive;
|
||
}
|
||
if (requestOptions.allowRetries != null) {
|
||
this._allowRetries = requestOptions.allowRetries;
|
||
}
|
||
if (requestOptions.maxRetries != null) {
|
||
this._maxRetries = requestOptions.maxRetries;
|
||
}
|
||
}
|
||
}
|
||
options(requestUrl, additionalHeaders) {
|
||
return this.request('OPTIONS', requestUrl, null, additionalHeaders || {});
|
||
}
|
||
get(requestUrl, additionalHeaders) {
|
||
return this.request('GET', requestUrl, null, additionalHeaders || {});
|
||
}
|
||
del(requestUrl, additionalHeaders) {
|
||
return this.request('DELETE', requestUrl, null, additionalHeaders || {});
|
||
}
|
||
post(requestUrl, data, additionalHeaders) {
|
||
return this.request('POST', requestUrl, data, additionalHeaders || {});
|
||
}
|
||
patch(requestUrl, data, additionalHeaders) {
|
||
return this.request('PATCH', requestUrl, data, additionalHeaders || {});
|
||
}
|
||
put(requestUrl, data, additionalHeaders) {
|
||
return this.request('PUT', requestUrl, data, additionalHeaders || {});
|
||
}
|
||
head(requestUrl, additionalHeaders) {
|
||
return this.request('HEAD', requestUrl, null, additionalHeaders || {});
|
||
}
|
||
sendStream(verb, requestUrl, stream, additionalHeaders) {
|
||
return this.request(verb, requestUrl, stream, additionalHeaders);
|
||
}
|
||
/**
|
||
* Gets a typed object from an endpoint
|
||
* Be aware that not found returns a null. Other errors (4xx, 5xx) reject the promise
|
||
*/
|
||
async getJson(requestUrl, additionalHeaders = {}) {
|
||
additionalHeaders[Headers.Accept] = this._getExistingOrDefaultHeader(additionalHeaders, Headers.Accept, MediaTypes.ApplicationJson);
|
||
let res = await this.get(requestUrl, additionalHeaders);
|
||
return this._processResponse(res, this.requestOptions);
|
||
}
|
||
async postJson(requestUrl, obj, additionalHeaders = {}) {
|
||
let data = JSON.stringify(obj, null, 2);
|
||
additionalHeaders[Headers.Accept] = this._getExistingOrDefaultHeader(additionalHeaders, Headers.Accept, MediaTypes.ApplicationJson);
|
||
additionalHeaders[Headers.ContentType] = this._getExistingOrDefaultHeader(additionalHeaders, Headers.ContentType, MediaTypes.ApplicationJson);
|
||
let res = await this.post(requestUrl, data, additionalHeaders);
|
||
return this._processResponse(res, this.requestOptions);
|
||
}
|
||
async putJson(requestUrl, obj, additionalHeaders = {}) {
|
||
let data = JSON.stringify(obj, null, 2);
|
||
additionalHeaders[Headers.Accept] = this._getExistingOrDefaultHeader(additionalHeaders, Headers.Accept, MediaTypes.ApplicationJson);
|
||
additionalHeaders[Headers.ContentType] = this._getExistingOrDefaultHeader(additionalHeaders, Headers.ContentType, MediaTypes.ApplicationJson);
|
||
let res = await this.put(requestUrl, data, additionalHeaders);
|
||
return this._processResponse(res, this.requestOptions);
|
||
}
|
||
async patchJson(requestUrl, obj, additionalHeaders = {}) {
|
||
let data = JSON.stringify(obj, null, 2);
|
||
additionalHeaders[Headers.Accept] = this._getExistingOrDefaultHeader(additionalHeaders, Headers.Accept, MediaTypes.ApplicationJson);
|
||
additionalHeaders[Headers.ContentType] = this._getExistingOrDefaultHeader(additionalHeaders, Headers.ContentType, MediaTypes.ApplicationJson);
|
||
let res = await this.patch(requestUrl, data, additionalHeaders);
|
||
return this._processResponse(res, this.requestOptions);
|
||
}
|
||
/**
|
||
* Makes a raw http request.
|
||
* All other methods such as get, post, patch, and request ultimately call this.
|
||
* Prefer get, del, post and patch
|
||
*/
|
||
async request(verb, requestUrl, data, headers) {
|
||
if (this._disposed) {
|
||
throw new Error('Client has already been disposed.');
|
||
}
|
||
let parsedUrl = new URL(requestUrl);
|
||
let info = this._prepareRequest(verb, parsedUrl, headers);
|
||
// Only perform retries on reads since writes may not be idempotent.
|
||
let maxTries = this._allowRetries && RetryableHttpVerbs.indexOf(verb) != -1
|
||
? this._maxRetries + 1
|
||
: 1;
|
||
let numTries = 0;
|
||
let response;
|
||
while (numTries < maxTries) {
|
||
response = await this.requestRaw(info, data);
|
||
// Check if it's an authentication challenge
|
||
if (response &&
|
||
response.message &&
|
||
response.message.statusCode === HttpCodes.Unauthorized) {
|
||
let authenticationHandler;
|
||
for (let i = 0; i < this.handlers.length; i++) {
|
||
if (this.handlers[i].canHandleAuthentication(response)) {
|
||
authenticationHandler = this.handlers[i];
|
||
break;
|
||
}
|
||
}
|
||
if (authenticationHandler) {
|
||
return authenticationHandler.handleAuthentication(this, info, data);
|
||
}
|
||
else {
|
||
// We have received an unauthorized response but have no handlers to handle it.
|
||
// Let the response return to the caller.
|
||
return response;
|
||
}
|
||
}
|
||
let redirectsRemaining = this._maxRedirects;
|
||
while (HttpRedirectCodes.indexOf(response.message.statusCode) != -1 &&
|
||
this._allowRedirects &&
|
||
redirectsRemaining > 0) {
|
||
const redirectUrl = response.message.headers['location'];
|
||
if (!redirectUrl) {
|
||
// if there's no location to redirect to, we won't
|
||
break;
|
||
}
|
||
let parsedRedirectUrl = new URL(redirectUrl);
|
||
if (parsedUrl.protocol == 'https:' &&
|
||
parsedUrl.protocol != parsedRedirectUrl.protocol &&
|
||
!this._allowRedirectDowngrade) {
|
||
throw new Error('Redirect from HTTPS to HTTP protocol. This downgrade is not allowed for security reasons. If you want to allow this behavior, set the allowRedirectDowngrade option to true.');
|
||
}
|
||
// we need to finish reading the response before reassigning response
|
||
// which will leak the open socket.
|
||
await response.readBody();
|
||
// strip authorization header if redirected to a different hostname
|
||
if (parsedRedirectUrl.hostname !== parsedUrl.hostname) {
|
||
for (let header in headers) {
|
||
// header names are case insensitive
|
||
if (header.toLowerCase() === 'authorization') {
|
||
delete headers[header];
|
||
}
|
||
}
|
||
}
|
||
// let's make the request with the new redirectUrl
|
||
info = this._prepareRequest(verb, parsedRedirectUrl, headers);
|
||
response = await this.requestRaw(info, data);
|
||
redirectsRemaining--;
|
||
}
|
||
if (HttpResponseRetryCodes.indexOf(response.message.statusCode) == -1) {
|
||
// If not a retry code, return immediately instead of retrying
|
||
return response;
|
||
}
|
||
numTries += 1;
|
||
if (numTries < maxTries) {
|
||
await response.readBody();
|
||
await this._performExponentialBackoff(numTries);
|
||
}
|
||
}
|
||
return response;
|
||
}
|
||
/**
|
||
* Needs to be called if keepAlive is set to true in request options.
|
||
*/
|
||
dispose() {
|
||
if (this._agent) {
|
||
this._agent.destroy();
|
||
}
|
||
this._disposed = true;
|
||
}
|
||
/**
|
||
* Raw request.
|
||
* @param info
|
||
* @param data
|
||
*/
|
||
requestRaw(info, data) {
|
||
return new Promise((resolve, reject) => {
|
||
let callbackForResult = function (err, res) {
|
||
if (err) {
|
||
reject(err);
|
||
}
|
||
resolve(res);
|
||
};
|
||
this.requestRawWithCallback(info, data, callbackForResult);
|
||
});
|
||
}
|
||
/**
|
||
* Raw request with callback.
|
||
* @param info
|
||
* @param data
|
||
* @param onResult
|
||
*/
|
||
requestRawWithCallback(info, data, onResult) {
|
||
let socket;
|
||
if (typeof data === 'string') {
|
||
info.options.headers['Content-Length'] = Buffer.byteLength(data, 'utf8');
|
||
}
|
||
let callbackCalled = false;
|
||
let handleResult = (err, res) => {
|
||
if (!callbackCalled) {
|
||
callbackCalled = true;
|
||
onResult(err, res);
|
||
}
|
||
};
|
||
let req = info.httpModule.request(info.options, (msg) => {
|
||
let res = new HttpClientResponse(msg);
|
||
handleResult(null, res);
|
||
});
|
||
req.on('socket', sock => {
|
||
socket = sock;
|
||
});
|
||
// If we ever get disconnected, we want the socket to timeout eventually
|
||
req.setTimeout(this._socketTimeout || 3 * 60000, () => {
|
||
if (socket) {
|
||
socket.end();
|
||
}
|
||
handleResult(new Error('Request timeout: ' + info.options.path), null);
|
||
});
|
||
req.on('error', function (err) {
|
||
// err has statusCode property
|
||
// res should have headers
|
||
handleResult(err, null);
|
||
});
|
||
if (data && typeof data === 'string') {
|
||
req.write(data, 'utf8');
|
||
}
|
||
if (data && typeof data !== 'string') {
|
||
data.on('close', function () {
|
||
req.end();
|
||
});
|
||
data.pipe(req);
|
||
}
|
||
else {
|
||
req.end();
|
||
}
|
||
}
|
||
/**
|
||
* Gets an http agent. This function is useful when you need an http agent that handles
|
||
* routing through a proxy server - depending upon the url and proxy environment variables.
|
||
* @param serverUrl The server URL where the request will be sent. For example, https://api.github.com
|
||
*/
|
||
getAgent(serverUrl) {
|
||
let parsedUrl = new URL(serverUrl);
|
||
return this._getAgent(parsedUrl);
|
||
}
|
||
_prepareRequest(method, requestUrl, headers) {
|
||
const info = {};
|
||
info.parsedUrl = requestUrl;
|
||
const usingSsl = info.parsedUrl.protocol === 'https:';
|
||
info.httpModule = usingSsl ? https : http;
|
||
const defaultPort = usingSsl ? 443 : 80;
|
||
info.options = {};
|
||
info.options.host = info.parsedUrl.hostname;
|
||
info.options.port = info.parsedUrl.port
|
||
? parseInt(info.parsedUrl.port)
|
||
: defaultPort;
|
||
info.options.path =
|
||
(info.parsedUrl.pathname || '') + (info.parsedUrl.search || '');
|
||
info.options.method = method;
|
||
info.options.headers = this._mergeHeaders(headers);
|
||
if (this.userAgent != null) {
|
||
info.options.headers['user-agent'] = this.userAgent;
|
||
}
|
||
info.options.agent = this._getAgent(info.parsedUrl);
|
||
// gives handlers an opportunity to participate
|
||
if (this.handlers) {
|
||
this.handlers.forEach(handler => {
|
||
handler.prepareRequest(info.options);
|
||
});
|
||
}
|
||
return info;
|
||
}
|
||
_mergeHeaders(headers) {
|
||
const lowercaseKeys = obj => Object.keys(obj).reduce((c, k) => ((c[k.toLowerCase()] = obj[k]), c), {});
|
||
if (this.requestOptions && this.requestOptions.headers) {
|
||
return Object.assign({}, lowercaseKeys(this.requestOptions.headers), lowercaseKeys(headers));
|
||
}
|
||
return lowercaseKeys(headers || {});
|
||
}
|
||
_getExistingOrDefaultHeader(additionalHeaders, header, _default) {
|
||
const lowercaseKeys = obj => Object.keys(obj).reduce((c, k) => ((c[k.toLowerCase()] = obj[k]), c), {});
|
||
let clientHeader;
|
||
if (this.requestOptions && this.requestOptions.headers) {
|
||
clientHeader = lowercaseKeys(this.requestOptions.headers)[header];
|
||
}
|
||
return additionalHeaders[header] || clientHeader || _default;
|
||
}
|
||
_getAgent(parsedUrl) {
|
||
let agent;
|
||
let proxyUrl = pm.getProxyUrl(parsedUrl);
|
||
let useProxy = proxyUrl && proxyUrl.hostname;
|
||
if (this._keepAlive && useProxy) {
|
||
agent = this._proxyAgent;
|
||
}
|
||
if (this._keepAlive && !useProxy) {
|
||
agent = this._agent;
|
||
}
|
||
// if agent is already assigned use that agent.
|
||
if (!!agent) {
|
||
return agent;
|
||
}
|
||
const usingSsl = parsedUrl.protocol === 'https:';
|
||
let maxSockets = 100;
|
||
if (!!this.requestOptions) {
|
||
maxSockets = this.requestOptions.maxSockets || http.globalAgent.maxSockets;
|
||
}
|
||
if (useProxy) {
|
||
// If using proxy, need tunnel
|
||
if (!tunnel) {
|
||
tunnel = __webpack_require__(856);
|
||
}
|
||
const agentOptions = {
|
||
maxSockets: maxSockets,
|
||
keepAlive: this._keepAlive,
|
||
proxy: {
|
||
...((proxyUrl.username || proxyUrl.password) && {
|
||
proxyAuth: `${proxyUrl.username}:${proxyUrl.password}`
|
||
}),
|
||
host: proxyUrl.hostname,
|
||
port: proxyUrl.port
|
||
}
|
||
};
|
||
let tunnelAgent;
|
||
const overHttps = proxyUrl.protocol === 'https:';
|
||
if (usingSsl) {
|
||
tunnelAgent = overHttps ? tunnel.httpsOverHttps : tunnel.httpsOverHttp;
|
||
}
|
||
else {
|
||
tunnelAgent = overHttps ? tunnel.httpOverHttps : tunnel.httpOverHttp;
|
||
}
|
||
agent = tunnelAgent(agentOptions);
|
||
this._proxyAgent = agent;
|
||
}
|
||
// if reusing agent across request and tunneling agent isn't assigned create a new agent
|
||
if (this._keepAlive && !agent) {
|
||
const options = { keepAlive: this._keepAlive, maxSockets: maxSockets };
|
||
agent = usingSsl ? new https.Agent(options) : new http.Agent(options);
|
||
this._agent = agent;
|
||
}
|
||
// if not using private agent and tunnel agent isn't setup then use global agent
|
||
if (!agent) {
|
||
agent = usingSsl ? https.globalAgent : http.globalAgent;
|
||
}
|
||
if (usingSsl && this._ignoreSslError) {
|
||
// we don't want to set NODE_TLS_REJECT_UNAUTHORIZED=0 since that will affect request for entire process
|
||
// http.RequestOptions doesn't expose a way to modify RequestOptions.agent.options
|
||
// we have to cast it to any and change it directly
|
||
agent.options = Object.assign(agent.options || {}, {
|
||
rejectUnauthorized: false
|
||
});
|
||
}
|
||
return agent;
|
||
}
|
||
_performExponentialBackoff(retryNumber) {
|
||
retryNumber = Math.min(ExponentialBackoffCeiling, retryNumber);
|
||
const ms = ExponentialBackoffTimeSlice * Math.pow(2, retryNumber);
|
||
return new Promise(resolve => setTimeout(() => resolve(), ms));
|
||
}
|
||
static dateTimeDeserializer(key, value) {
|
||
if (typeof value === 'string') {
|
||
let a = new Date(value);
|
||
if (!isNaN(a.valueOf())) {
|
||
return a;
|
||
}
|
||
}
|
||
return value;
|
||
}
|
||
async _processResponse(res, options) {
|
||
return new Promise(async (resolve, reject) => {
|
||
const statusCode = res.message.statusCode;
|
||
const response = {
|
||
statusCode: statusCode,
|
||
result: null,
|
||
headers: {}
|
||
};
|
||
// not found leads to null obj returned
|
||
if (statusCode == HttpCodes.NotFound) {
|
||
resolve(response);
|
||
}
|
||
let obj;
|
||
let contents;
|
||
// get the result from the body
|
||
try {
|
||
contents = await res.readBody();
|
||
if (contents && contents.length > 0) {
|
||
if (options && options.deserializeDates) {
|
||
obj = JSON.parse(contents, HttpClient.dateTimeDeserializer);
|
||
}
|
||
else {
|
||
obj = JSON.parse(contents);
|
||
}
|
||
response.result = obj;
|
||
}
|
||
response.headers = res.message.headers;
|
||
}
|
||
catch (err) {
|
||
// Invalid resource (contents not json); leaving result obj null
|
||
}
|
||
// note that 3xx redirects are handled by the http layer.
|
||
if (statusCode > 299) {
|
||
let msg;
|
||
// if exception/error in body, attempt to get better error
|
||
if (obj && obj.message) {
|
||
msg = obj.message;
|
||
}
|
||
else if (contents && contents.length > 0) {
|
||
// it may be the case that the exception is in the body message as string
|
||
msg = contents;
|
||
}
|
||
else {
|
||
msg = 'Failed request: (' + statusCode + ')';
|
||
}
|
||
let err = new HttpClientError(msg, statusCode);
|
||
err.result = response.result;
|
||
reject(err);
|
||
}
|
||
else {
|
||
resolve(response);
|
||
}
|
||
});
|
||
}
|
||
}
|
||
exports.HttpClient = HttpClient;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 549:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
var url = __webpack_require__(835);
|
||
var URL = url.URL;
|
||
var http = __webpack_require__(605);
|
||
var https = __webpack_require__(211);
|
||
var Writable = __webpack_require__(413).Writable;
|
||
var assert = __webpack_require__(357);
|
||
var debug = __webpack_require__(454);
|
||
|
||
// Create handlers that pass events from native requests
|
||
var events = ["abort", "aborted", "connect", "error", "socket", "timeout"];
|
||
var eventHandlers = Object.create(null);
|
||
events.forEach(function (event) {
|
||
eventHandlers[event] = function (arg1, arg2, arg3) {
|
||
this._redirectable.emit(event, arg1, arg2, arg3);
|
||
};
|
||
});
|
||
|
||
// Error types with codes
|
||
var RedirectionError = createErrorType(
|
||
"ERR_FR_REDIRECTION_FAILURE",
|
||
""
|
||
);
|
||
var TooManyRedirectsError = createErrorType(
|
||
"ERR_FR_TOO_MANY_REDIRECTS",
|
||
"Maximum number of redirects exceeded"
|
||
);
|
||
var MaxBodyLengthExceededError = createErrorType(
|
||
"ERR_FR_MAX_BODY_LENGTH_EXCEEDED",
|
||
"Request body larger than maxBodyLength limit"
|
||
);
|
||
var WriteAfterEndError = createErrorType(
|
||
"ERR_STREAM_WRITE_AFTER_END",
|
||
"write after end"
|
||
);
|
||
|
||
// An HTTP(S) request that can be redirected
|
||
function RedirectableRequest(options, responseCallback) {
|
||
// Initialize the request
|
||
Writable.call(this);
|
||
this._sanitizeOptions(options);
|
||
this._options = options;
|
||
this._ended = false;
|
||
this._ending = false;
|
||
this._redirectCount = 0;
|
||
this._redirects = [];
|
||
this._requestBodyLength = 0;
|
||
this._requestBodyBuffers = [];
|
||
|
||
// Attach a callback if passed
|
||
if (responseCallback) {
|
||
this.on("response", responseCallback);
|
||
}
|
||
|
||
// React to responses of native requests
|
||
var self = this;
|
||
this._onNativeResponse = function (response) {
|
||
self._processResponse(response);
|
||
};
|
||
|
||
// Perform the first request
|
||
this._performRequest();
|
||
}
|
||
RedirectableRequest.prototype = Object.create(Writable.prototype);
|
||
|
||
RedirectableRequest.prototype.abort = function () {
|
||
abortRequest(this._currentRequest);
|
||
this.emit("abort");
|
||
};
|
||
|
||
// Writes buffered data to the current native request
|
||
RedirectableRequest.prototype.write = function (data, encoding, callback) {
|
||
// Writing is not allowed if end has been called
|
||
if (this._ending) {
|
||
throw new WriteAfterEndError();
|
||
}
|
||
|
||
// Validate input and shift parameters if necessary
|
||
if (!(typeof data === "string" || typeof data === "object" && ("length" in data))) {
|
||
throw new TypeError("data should be a string, Buffer or Uint8Array");
|
||
}
|
||
if (typeof encoding === "function") {
|
||
callback = encoding;
|
||
encoding = null;
|
||
}
|
||
|
||
// Ignore empty buffers, since writing them doesn't invoke the callback
|
||
// https://github.com/nodejs/node/issues/22066
|
||
if (data.length === 0) {
|
||
if (callback) {
|
||
callback();
|
||
}
|
||
return;
|
||
}
|
||
// Only write when we don't exceed the maximum body length
|
||
if (this._requestBodyLength + data.length <= this._options.maxBodyLength) {
|
||
this._requestBodyLength += data.length;
|
||
this._requestBodyBuffers.push({ data: data, encoding: encoding });
|
||
this._currentRequest.write(data, encoding, callback);
|
||
}
|
||
// Error when we exceed the maximum body length
|
||
else {
|
||
this.emit("error", new MaxBodyLengthExceededError());
|
||
this.abort();
|
||
}
|
||
};
|
||
|
||
// Ends the current native request
|
||
RedirectableRequest.prototype.end = function (data, encoding, callback) {
|
||
// Shift parameters if necessary
|
||
if (typeof data === "function") {
|
||
callback = data;
|
||
data = encoding = null;
|
||
}
|
||
else if (typeof encoding === "function") {
|
||
callback = encoding;
|
||
encoding = null;
|
||
}
|
||
|
||
// Write data if needed and end
|
||
if (!data) {
|
||
this._ended = this._ending = true;
|
||
this._currentRequest.end(null, null, callback);
|
||
}
|
||
else {
|
||
var self = this;
|
||
var currentRequest = this._currentRequest;
|
||
this.write(data, encoding, function () {
|
||
self._ended = true;
|
||
currentRequest.end(null, null, callback);
|
||
});
|
||
this._ending = true;
|
||
}
|
||
};
|
||
|
||
// Sets a header value on the current native request
|
||
RedirectableRequest.prototype.setHeader = function (name, value) {
|
||
this._options.headers[name] = value;
|
||
this._currentRequest.setHeader(name, value);
|
||
};
|
||
|
||
// Clears a header value on the current native request
|
||
RedirectableRequest.prototype.removeHeader = function (name) {
|
||
delete this._options.headers[name];
|
||
this._currentRequest.removeHeader(name);
|
||
};
|
||
|
||
// Global timeout for all underlying requests
|
||
RedirectableRequest.prototype.setTimeout = function (msecs, callback) {
|
||
var self = this;
|
||
|
||
// Destroys the socket on timeout
|
||
function destroyOnTimeout(socket) {
|
||
socket.setTimeout(msecs);
|
||
socket.removeListener("timeout", socket.destroy);
|
||
socket.addListener("timeout", socket.destroy);
|
||
}
|
||
|
||
// Sets up a timer to trigger a timeout event
|
||
function startTimer(socket) {
|
||
if (self._timeout) {
|
||
clearTimeout(self._timeout);
|
||
}
|
||
self._timeout = setTimeout(function () {
|
||
self.emit("timeout");
|
||
clearTimer();
|
||
}, msecs);
|
||
destroyOnTimeout(socket);
|
||
}
|
||
|
||
// Stops a timeout from triggering
|
||
function clearTimer() {
|
||
if (self._timeout) {
|
||
clearTimeout(self._timeout);
|
||
self._timeout = null;
|
||
}
|
||
if (callback) {
|
||
self.removeListener("timeout", callback);
|
||
}
|
||
if (!self.socket) {
|
||
self._currentRequest.removeListener("socket", startTimer);
|
||
}
|
||
}
|
||
|
||
// Attach callback if passed
|
||
if (callback) {
|
||
this.on("timeout", callback);
|
||
}
|
||
|
||
// Start the timer if or when the socket is opened
|
||
if (this.socket) {
|
||
startTimer(this.socket);
|
||
}
|
||
else {
|
||
this._currentRequest.once("socket", startTimer);
|
||
}
|
||
|
||
// Clean up on events
|
||
this.on("socket", destroyOnTimeout);
|
||
this.once("response", clearTimer);
|
||
this.once("error", clearTimer);
|
||
|
||
return this;
|
||
};
|
||
|
||
// Proxy all other public ClientRequest methods
|
||
[
|
||
"flushHeaders", "getHeader",
|
||
"setNoDelay", "setSocketKeepAlive",
|
||
].forEach(function (method) {
|
||
RedirectableRequest.prototype[method] = function (a, b) {
|
||
return this._currentRequest[method](a, b);
|
||
};
|
||
});
|
||
|
||
// Proxy all public ClientRequest properties
|
||
["aborted", "connection", "socket"].forEach(function (property) {
|
||
Object.defineProperty(RedirectableRequest.prototype, property, {
|
||
get: function () { return this._currentRequest[property]; },
|
||
});
|
||
});
|
||
|
||
RedirectableRequest.prototype._sanitizeOptions = function (options) {
|
||
// Ensure headers are always present
|
||
if (!options.headers) {
|
||
options.headers = {};
|
||
}
|
||
|
||
// Since http.request treats host as an alias of hostname,
|
||
// but the url module interprets host as hostname plus port,
|
||
// eliminate the host property to avoid confusion.
|
||
if (options.host) {
|
||
// Use hostname if set, because it has precedence
|
||
if (!options.hostname) {
|
||
options.hostname = options.host;
|
||
}
|
||
delete options.host;
|
||
}
|
||
|
||
// Complete the URL object when necessary
|
||
if (!options.pathname && options.path) {
|
||
var searchPos = options.path.indexOf("?");
|
||
if (searchPos < 0) {
|
||
options.pathname = options.path;
|
||
}
|
||
else {
|
||
options.pathname = options.path.substring(0, searchPos);
|
||
options.search = options.path.substring(searchPos);
|
||
}
|
||
}
|
||
};
|
||
|
||
|
||
// Executes the next native request (initial or redirect)
|
||
RedirectableRequest.prototype._performRequest = function () {
|
||
// Load the native protocol
|
||
var protocol = this._options.protocol;
|
||
var nativeProtocol = this._options.nativeProtocols[protocol];
|
||
if (!nativeProtocol) {
|
||
this.emit("error", new TypeError("Unsupported protocol " + protocol));
|
||
return;
|
||
}
|
||
|
||
// If specified, use the agent corresponding to the protocol
|
||
// (HTTP and HTTPS use different types of agents)
|
||
if (this._options.agents) {
|
||
var scheme = protocol.substr(0, protocol.length - 1);
|
||
this._options.agent = this._options.agents[scheme];
|
||
}
|
||
|
||
// Create the native request
|
||
var request = this._currentRequest =
|
||
nativeProtocol.request(this._options, this._onNativeResponse);
|
||
this._currentUrl = url.format(this._options);
|
||
|
||
// Set up event handlers
|
||
request._redirectable = this;
|
||
for (var e = 0; e < events.length; e++) {
|
||
request.on(events[e], eventHandlers[events[e]]);
|
||
}
|
||
|
||
// End a redirected request
|
||
// (The first request must be ended explicitly with RedirectableRequest#end)
|
||
if (this._isRedirect) {
|
||
// Write the request entity and end.
|
||
var i = 0;
|
||
var self = this;
|
||
var buffers = this._requestBodyBuffers;
|
||
(function writeNext(error) {
|
||
// Only write if this request has not been redirected yet
|
||
/* istanbul ignore else */
|
||
if (request === self._currentRequest) {
|
||
// Report any write errors
|
||
/* istanbul ignore if */
|
||
if (error) {
|
||
self.emit("error", error);
|
||
}
|
||
// Write the next buffer if there are still left
|
||
else if (i < buffers.length) {
|
||
var buffer = buffers[i++];
|
||
/* istanbul ignore else */
|
||
if (!request.finished) {
|
||
request.write(buffer.data, buffer.encoding, writeNext);
|
||
}
|
||
}
|
||
// End the request if `end` has been called on us
|
||
else if (self._ended) {
|
||
request.end();
|
||
}
|
||
}
|
||
}());
|
||
}
|
||
};
|
||
|
||
// Processes a response from the current native request
|
||
RedirectableRequest.prototype._processResponse = function (response) {
|
||
// Store the redirected response
|
||
var statusCode = response.statusCode;
|
||
if (this._options.trackRedirects) {
|
||
this._redirects.push({
|
||
url: this._currentUrl,
|
||
headers: response.headers,
|
||
statusCode: statusCode,
|
||
});
|
||
}
|
||
|
||
// RFC7231§6.4: The 3xx (Redirection) class of status code indicates
|
||
// that further action needs to be taken by the user agent in order to
|
||
// fulfill the request. If a Location header field is provided,
|
||
// the user agent MAY automatically redirect its request to the URI
|
||
// referenced by the Location field value,
|
||
// even if the specific status code is not understood.
|
||
var location = response.headers.location;
|
||
if (location && this._options.followRedirects !== false &&
|
||
statusCode >= 300 && statusCode < 400) {
|
||
// Abort the current request
|
||
abortRequest(this._currentRequest);
|
||
// Discard the remainder of the response to avoid waiting for data
|
||
response.destroy();
|
||
|
||
// RFC7231§6.4: A client SHOULD detect and intervene
|
||
// in cyclical redirections (i.e., "infinite" redirection loops).
|
||
if (++this._redirectCount > this._options.maxRedirects) {
|
||
this.emit("error", new TooManyRedirectsError());
|
||
return;
|
||
}
|
||
|
||
// RFC7231§6.4: Automatic redirection needs to done with
|
||
// care for methods not known to be safe, […]
|
||
// RFC7231§6.4.2–3: For historical reasons, a user agent MAY change
|
||
// the request method from POST to GET for the subsequent request.
|
||
if ((statusCode === 301 || statusCode === 302) && this._options.method === "POST" ||
|
||
// RFC7231§6.4.4: The 303 (See Other) status code indicates that
|
||
// the server is redirecting the user agent to a different resource […]
|
||
// A user agent can perform a retrieval request targeting that URI
|
||
// (a GET or HEAD request if using HTTP) […]
|
||
(statusCode === 303) && !/^(?:GET|HEAD)$/.test(this._options.method)) {
|
||
this._options.method = "GET";
|
||
// Drop a possible entity and headers related to it
|
||
this._requestBodyBuffers = [];
|
||
removeMatchingHeaders(/^content-/i, this._options.headers);
|
||
}
|
||
|
||
// Drop the Host header, as the redirect might lead to a different host
|
||
var previousHostName = removeMatchingHeaders(/^host$/i, this._options.headers) ||
|
||
url.parse(this._currentUrl).hostname;
|
||
|
||
// Create the redirected request
|
||
var redirectUrl = url.resolve(this._currentUrl, location);
|
||
debug("redirecting to", redirectUrl);
|
||
this._isRedirect = true;
|
||
var redirectUrlParts = url.parse(redirectUrl);
|
||
Object.assign(this._options, redirectUrlParts);
|
||
|
||
// Drop the Authorization header if redirecting to another host
|
||
if (redirectUrlParts.hostname !== previousHostName) {
|
||
removeMatchingHeaders(/^authorization$/i, this._options.headers);
|
||
}
|
||
|
||
// Evaluate the beforeRedirect callback
|
||
if (typeof this._options.beforeRedirect === "function") {
|
||
var responseDetails = { headers: response.headers };
|
||
try {
|
||
this._options.beforeRedirect.call(null, this._options, responseDetails);
|
||
}
|
||
catch (err) {
|
||
this.emit("error", err);
|
||
return;
|
||
}
|
||
this._sanitizeOptions(this._options);
|
||
}
|
||
|
||
// Perform the redirected request
|
||
try {
|
||
this._performRequest();
|
||
}
|
||
catch (cause) {
|
||
var error = new RedirectionError("Redirected request failed: " + cause.message);
|
||
error.cause = cause;
|
||
this.emit("error", error);
|
||
}
|
||
}
|
||
else {
|
||
// The response is not a redirect; return it as-is
|
||
response.responseUrl = this._currentUrl;
|
||
response.redirects = this._redirects;
|
||
this.emit("response", response);
|
||
|
||
// Clean up
|
||
this._requestBodyBuffers = [];
|
||
}
|
||
};
|
||
|
||
// Wraps the key/value object of protocols with redirect functionality
|
||
function wrap(protocols) {
|
||
// Default settings
|
||
var exports = {
|
||
maxRedirects: 21,
|
||
maxBodyLength: 10 * 1024 * 1024,
|
||
};
|
||
|
||
// Wrap each protocol
|
||
var nativeProtocols = {};
|
||
Object.keys(protocols).forEach(function (scheme) {
|
||
var protocol = scheme + ":";
|
||
var nativeProtocol = nativeProtocols[protocol] = protocols[scheme];
|
||
var wrappedProtocol = exports[scheme] = Object.create(nativeProtocol);
|
||
|
||
// Executes a request, following redirects
|
||
function request(input, options, callback) {
|
||
// Parse parameters
|
||
if (typeof input === "string") {
|
||
var urlStr = input;
|
||
try {
|
||
input = urlToOptions(new URL(urlStr));
|
||
}
|
||
catch (err) {
|
||
/* istanbul ignore next */
|
||
input = url.parse(urlStr);
|
||
}
|
||
}
|
||
else if (URL && (input instanceof URL)) {
|
||
input = urlToOptions(input);
|
||
}
|
||
else {
|
||
callback = options;
|
||
options = input;
|
||
input = { protocol: protocol };
|
||
}
|
||
if (typeof options === "function") {
|
||
callback = options;
|
||
options = null;
|
||
}
|
||
|
||
// Set defaults
|
||
options = Object.assign({
|
||
maxRedirects: exports.maxRedirects,
|
||
maxBodyLength: exports.maxBodyLength,
|
||
}, input, options);
|
||
options.nativeProtocols = nativeProtocols;
|
||
|
||
assert.equal(options.protocol, protocol, "protocol mismatch");
|
||
debug("options", options);
|
||
return new RedirectableRequest(options, callback);
|
||
}
|
||
|
||
// Executes a GET request, following redirects
|
||
function get(input, options, callback) {
|
||
var wrappedRequest = wrappedProtocol.request(input, options, callback);
|
||
wrappedRequest.end();
|
||
return wrappedRequest;
|
||
}
|
||
|
||
// Expose the properties on the wrapped protocol
|
||
Object.defineProperties(wrappedProtocol, {
|
||
request: { value: request, configurable: true, enumerable: true, writable: true },
|
||
get: { value: get, configurable: true, enumerable: true, writable: true },
|
||
});
|
||
});
|
||
return exports;
|
||
}
|
||
|
||
/* istanbul ignore next */
|
||
function noop() { /* empty */ }
|
||
|
||
// from https://github.com/nodejs/node/blob/master/lib/internal/url.js
|
||
function urlToOptions(urlObject) {
|
||
var options = {
|
||
protocol: urlObject.protocol,
|
||
hostname: urlObject.hostname.startsWith("[") ?
|
||
/* istanbul ignore next */
|
||
urlObject.hostname.slice(1, -1) :
|
||
urlObject.hostname,
|
||
hash: urlObject.hash,
|
||
search: urlObject.search,
|
||
pathname: urlObject.pathname,
|
||
path: urlObject.pathname + urlObject.search,
|
||
href: urlObject.href,
|
||
};
|
||
if (urlObject.port !== "") {
|
||
options.port = Number(urlObject.port);
|
||
}
|
||
return options;
|
||
}
|
||
|
||
function removeMatchingHeaders(regex, headers) {
|
||
var lastValue;
|
||
for (var header in headers) {
|
||
if (regex.test(header)) {
|
||
lastValue = headers[header];
|
||
delete headers[header];
|
||
}
|
||
}
|
||
return lastValue;
|
||
}
|
||
|
||
function createErrorType(code, defaultMessage) {
|
||
function CustomError(message) {
|
||
Error.captureStackTrace(this, this.constructor);
|
||
this.message = message || defaultMessage;
|
||
}
|
||
CustomError.prototype = new Error();
|
||
CustomError.prototype.constructor = CustomError;
|
||
CustomError.prototype.name = "Error [" + code + "]";
|
||
CustomError.prototype.code = code;
|
||
return CustomError;
|
||
}
|
||
|
||
function abortRequest(request) {
|
||
for (var e = 0; e < events.length; e++) {
|
||
request.removeListener(events[e], eventHandlers[events[e]]);
|
||
}
|
||
request.on("error", noop);
|
||
request.abort();
|
||
}
|
||
|
||
// Exports
|
||
module.exports = wrap({ http: http, https: https });
|
||
module.exports.wrap = wrap;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 564:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var createError = __webpack_require__(26);
|
||
|
||
/**
|
||
* Resolve or reject a Promise based on response status.
|
||
*
|
||
* @param {Function} resolve A function that resolves the promise.
|
||
* @param {Function} reject A function that rejects the promise.
|
||
* @param {object} response The response.
|
||
*/
|
||
module.exports = function settle(resolve, reject, response) {
|
||
var validateStatus = response.config.validateStatus;
|
||
if (!response.status || !validateStatus || validateStatus(response.status)) {
|
||
resolve(response);
|
||
} else {
|
||
reject(createError(
|
||
'Request failed with status code ' + response.status,
|
||
response.config,
|
||
null,
|
||
response.request,
|
||
response
|
||
));
|
||
}
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 569:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
module.exports = rimraf
|
||
rimraf.sync = rimrafSync
|
||
|
||
var assert = __webpack_require__(357)
|
||
var path = __webpack_require__(622)
|
||
var fs = __webpack_require__(747)
|
||
var glob = __webpack_require__(402)
|
||
var _0666 = parseInt('666', 8)
|
||
|
||
var defaultGlobOpts = {
|
||
nosort: true,
|
||
silent: true
|
||
}
|
||
|
||
// for EMFILE handling
|
||
var timeout = 0
|
||
|
||
var isWindows = (process.platform === "win32")
|
||
|
||
function defaults (options) {
|
||
var methods = [
|
||
'unlink',
|
||
'chmod',
|
||
'stat',
|
||
'lstat',
|
||
'rmdir',
|
||
'readdir'
|
||
]
|
||
methods.forEach(function(m) {
|
||
options[m] = options[m] || fs[m]
|
||
m = m + 'Sync'
|
||
options[m] = options[m] || fs[m]
|
||
})
|
||
|
||
options.maxBusyTries = options.maxBusyTries || 3
|
||
options.emfileWait = options.emfileWait || 1000
|
||
if (options.glob === false) {
|
||
options.disableGlob = true
|
||
}
|
||
options.disableGlob = options.disableGlob || false
|
||
options.glob = options.glob || defaultGlobOpts
|
||
}
|
||
|
||
function rimraf (p, options, cb) {
|
||
if (typeof options === 'function') {
|
||
cb = options
|
||
options = {}
|
||
}
|
||
|
||
assert(p, 'rimraf: missing path')
|
||
assert.equal(typeof p, 'string', 'rimraf: path should be a string')
|
||
assert.equal(typeof cb, 'function', 'rimraf: callback function required')
|
||
assert(options, 'rimraf: invalid options argument provided')
|
||
assert.equal(typeof options, 'object', 'rimraf: options should be object')
|
||
|
||
defaults(options)
|
||
|
||
var busyTries = 0
|
||
var errState = null
|
||
var n = 0
|
||
|
||
if (options.disableGlob || !glob.hasMagic(p))
|
||
return afterGlob(null, [p])
|
||
|
||
options.lstat(p, function (er, stat) {
|
||
if (!er)
|
||
return afterGlob(null, [p])
|
||
|
||
glob(p, options.glob, afterGlob)
|
||
})
|
||
|
||
function next (er) {
|
||
errState = errState || er
|
||
if (--n === 0)
|
||
cb(errState)
|
||
}
|
||
|
||
function afterGlob (er, results) {
|
||
if (er)
|
||
return cb(er)
|
||
|
||
n = results.length
|
||
if (n === 0)
|
||
return cb()
|
||
|
||
results.forEach(function (p) {
|
||
rimraf_(p, options, function CB (er) {
|
||
if (er) {
|
||
if ((er.code === "EBUSY" || er.code === "ENOTEMPTY" || er.code === "EPERM") &&
|
||
busyTries < options.maxBusyTries) {
|
||
busyTries ++
|
||
var time = busyTries * 100
|
||
// try again, with the same exact callback as this one.
|
||
return setTimeout(function () {
|
||
rimraf_(p, options, CB)
|
||
}, time)
|
||
}
|
||
|
||
// this one won't happen if graceful-fs is used.
|
||
if (er.code === "EMFILE" && timeout < options.emfileWait) {
|
||
return setTimeout(function () {
|
||
rimraf_(p, options, CB)
|
||
}, timeout ++)
|
||
}
|
||
|
||
// already gone
|
||
if (er.code === "ENOENT") er = null
|
||
}
|
||
|
||
timeout = 0
|
||
next(er)
|
||
})
|
||
})
|
||
}
|
||
}
|
||
|
||
// Two possible strategies.
|
||
// 1. Assume it's a file. unlink it, then do the dir stuff on EPERM or EISDIR
|
||
// 2. Assume it's a directory. readdir, then do the file stuff on ENOTDIR
|
||
//
|
||
// Both result in an extra syscall when you guess wrong. However, there
|
||
// are likely far more normal files in the world than directories. This
|
||
// is based on the assumption that a the average number of files per
|
||
// directory is >= 1.
|
||
//
|
||
// If anyone ever complains about this, then I guess the strategy could
|
||
// be made configurable somehow. But until then, YAGNI.
|
||
function rimraf_ (p, options, cb) {
|
||
assert(p)
|
||
assert(options)
|
||
assert(typeof cb === 'function')
|
||
|
||
// sunos lets the root user unlink directories, which is... weird.
|
||
// so we have to lstat here and make sure it's not a dir.
|
||
options.lstat(p, function (er, st) {
|
||
if (er && er.code === "ENOENT")
|
||
return cb(null)
|
||
|
||
// Windows can EPERM on stat. Life is suffering.
|
||
if (er && er.code === "EPERM" && isWindows)
|
||
fixWinEPERM(p, options, er, cb)
|
||
|
||
if (st && st.isDirectory())
|
||
return rmdir(p, options, er, cb)
|
||
|
||
options.unlink(p, function (er) {
|
||
if (er) {
|
||
if (er.code === "ENOENT")
|
||
return cb(null)
|
||
if (er.code === "EPERM")
|
||
return (isWindows)
|
||
? fixWinEPERM(p, options, er, cb)
|
||
: rmdir(p, options, er, cb)
|
||
if (er.code === "EISDIR")
|
||
return rmdir(p, options, er, cb)
|
||
}
|
||
return cb(er)
|
||
})
|
||
})
|
||
}
|
||
|
||
function fixWinEPERM (p, options, er, cb) {
|
||
assert(p)
|
||
assert(options)
|
||
assert(typeof cb === 'function')
|
||
if (er)
|
||
assert(er instanceof Error)
|
||
|
||
options.chmod(p, _0666, function (er2) {
|
||
if (er2)
|
||
cb(er2.code === "ENOENT" ? null : er)
|
||
else
|
||
options.stat(p, function(er3, stats) {
|
||
if (er3)
|
||
cb(er3.code === "ENOENT" ? null : er)
|
||
else if (stats.isDirectory())
|
||
rmdir(p, options, er, cb)
|
||
else
|
||
options.unlink(p, cb)
|
||
})
|
||
})
|
||
}
|
||
|
||
function fixWinEPERMSync (p, options, er) {
|
||
assert(p)
|
||
assert(options)
|
||
if (er)
|
||
assert(er instanceof Error)
|
||
|
||
try {
|
||
options.chmodSync(p, _0666)
|
||
} catch (er2) {
|
||
if (er2.code === "ENOENT")
|
||
return
|
||
else
|
||
throw er
|
||
}
|
||
|
||
try {
|
||
var stats = options.statSync(p)
|
||
} catch (er3) {
|
||
if (er3.code === "ENOENT")
|
||
return
|
||
else
|
||
throw er
|
||
}
|
||
|
||
if (stats.isDirectory())
|
||
rmdirSync(p, options, er)
|
||
else
|
||
options.unlinkSync(p)
|
||
}
|
||
|
||
function rmdir (p, options, originalEr, cb) {
|
||
assert(p)
|
||
assert(options)
|
||
if (originalEr)
|
||
assert(originalEr instanceof Error)
|
||
assert(typeof cb === 'function')
|
||
|
||
// try to rmdir first, and only readdir on ENOTEMPTY or EEXIST (SunOS)
|
||
// if we guessed wrong, and it's not a directory, then
|
||
// raise the original error.
|
||
options.rmdir(p, function (er) {
|
||
if (er && (er.code === "ENOTEMPTY" || er.code === "EEXIST" || er.code === "EPERM"))
|
||
rmkids(p, options, cb)
|
||
else if (er && er.code === "ENOTDIR")
|
||
cb(originalEr)
|
||
else
|
||
cb(er)
|
||
})
|
||
}
|
||
|
||
function rmkids(p, options, cb) {
|
||
assert(p)
|
||
assert(options)
|
||
assert(typeof cb === 'function')
|
||
|
||
options.readdir(p, function (er, files) {
|
||
if (er)
|
||
return cb(er)
|
||
var n = files.length
|
||
if (n === 0)
|
||
return options.rmdir(p, cb)
|
||
var errState
|
||
files.forEach(function (f) {
|
||
rimraf(path.join(p, f), options, function (er) {
|
||
if (errState)
|
||
return
|
||
if (er)
|
||
return cb(errState = er)
|
||
if (--n === 0)
|
||
options.rmdir(p, cb)
|
||
})
|
||
})
|
||
})
|
||
}
|
||
|
||
// this looks simpler, and is strictly *faster*, but will
|
||
// tie up the JavaScript thread and fail on excessively
|
||
// deep directory trees.
|
||
function rimrafSync (p, options) {
|
||
options = options || {}
|
||
defaults(options)
|
||
|
||
assert(p, 'rimraf: missing path')
|
||
assert.equal(typeof p, 'string', 'rimraf: path should be a string')
|
||
assert(options, 'rimraf: missing options')
|
||
assert.equal(typeof options, 'object', 'rimraf: options should be object')
|
||
|
||
var results
|
||
|
||
if (options.disableGlob || !glob.hasMagic(p)) {
|
||
results = [p]
|
||
} else {
|
||
try {
|
||
options.lstatSync(p)
|
||
results = [p]
|
||
} catch (er) {
|
||
results = glob.sync(p, options.glob)
|
||
}
|
||
}
|
||
|
||
if (!results.length)
|
||
return
|
||
|
||
for (var i = 0; i < results.length; i++) {
|
||
var p = results[i]
|
||
|
||
try {
|
||
var st = options.lstatSync(p)
|
||
} catch (er) {
|
||
if (er.code === "ENOENT")
|
||
return
|
||
|
||
// Windows can EPERM on stat. Life is suffering.
|
||
if (er.code === "EPERM" && isWindows)
|
||
fixWinEPERMSync(p, options, er)
|
||
}
|
||
|
||
try {
|
||
// sunos lets the root user unlink directories, which is... weird.
|
||
if (st && st.isDirectory())
|
||
rmdirSync(p, options, null)
|
||
else
|
||
options.unlinkSync(p)
|
||
} catch (er) {
|
||
if (er.code === "ENOENT")
|
||
return
|
||
if (er.code === "EPERM")
|
||
return isWindows ? fixWinEPERMSync(p, options, er) : rmdirSync(p, options, er)
|
||
if (er.code !== "EISDIR")
|
||
throw er
|
||
|
||
rmdirSync(p, options, er)
|
||
}
|
||
}
|
||
}
|
||
|
||
function rmdirSync (p, options, originalEr) {
|
||
assert(p)
|
||
assert(options)
|
||
if (originalEr)
|
||
assert(originalEr instanceof Error)
|
||
|
||
try {
|
||
options.rmdirSync(p)
|
||
} catch (er) {
|
||
if (er.code === "ENOENT")
|
||
return
|
||
if (er.code === "ENOTDIR")
|
||
throw originalEr
|
||
if (er.code === "ENOTEMPTY" || er.code === "EEXIST" || er.code === "EPERM")
|
||
rmkidsSync(p, options)
|
||
}
|
||
}
|
||
|
||
function rmkidsSync (p, options) {
|
||
assert(p)
|
||
assert(options)
|
||
options.readdirSync(p).forEach(function (f) {
|
||
rimrafSync(path.join(p, f), options)
|
||
})
|
||
|
||
// We only end up here once we got ENOTEMPTY at least once, and
|
||
// at this point, we are guaranteed to have removed all the kids.
|
||
// So, we know that it won't be ENOENT or ENOTDIR or anything else.
|
||
// try really hard to delete stuff on windows, because it has a
|
||
// PROFOUNDLY annoying habit of not closing handles promptly when
|
||
// files are deleted, resulting in spurious ENOTEMPTY errors.
|
||
var retries = isWindows ? 100 : 1
|
||
var i = 0
|
||
do {
|
||
var threw = true
|
||
try {
|
||
var ret = options.rmdirSync(p, options)
|
||
threw = false
|
||
return ret
|
||
} finally {
|
||
if (++i < retries && threw)
|
||
continue
|
||
}
|
||
} while (true)
|
||
}
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 575:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
||
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
||
return new (P || (P = Promise))(function (resolve, reject) {
|
||
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
||
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
||
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
||
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
||
});
|
||
};
|
||
var __importStar = (this && this.__importStar) || function (mod) {
|
||
if (mod && mod.__esModule) return mod;
|
||
var result = {};
|
||
if (mod != null) for (var k in mod) if (Object.hasOwnProperty.call(mod, k)) result[k] = mod[k];
|
||
result["default"] = mod;
|
||
return result;
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const glob = __importStar(__webpack_require__(281));
|
||
const path = __importStar(__webpack_require__(622));
|
||
const core_1 = __webpack_require__(470);
|
||
const fs_1 = __webpack_require__(747);
|
||
const path_1 = __webpack_require__(622);
|
||
const util_1 = __webpack_require__(669);
|
||
const stats = util_1.promisify(fs_1.stat);
|
||
function getDefaultGlobOptions() {
|
||
return {
|
||
followSymbolicLinks: true,
|
||
implicitDescendants: true,
|
||
omitBrokenSymbolicLinks: true
|
||
};
|
||
}
|
||
/**
|
||
* If multiple paths are specific, the least common ancestor (LCA) of the search paths is used as
|
||
* the delimiter to control the directory structure for the artifact. This function returns the LCA
|
||
* when given an array of search paths
|
||
*
|
||
* Example 1: The patterns `/foo/` and `/bar/` returns `/`
|
||
*
|
||
* Example 2: The patterns `~/foo/bar/*` and `~/foo/voo/two/*` and `~/foo/mo/` returns `~/foo`
|
||
*/
|
||
function getMultiPathLCA(searchPaths) {
|
||
if (searchPaths.length < 2) {
|
||
throw new Error('At least two search paths must be provided');
|
||
}
|
||
const commonPaths = new Array();
|
||
const splitPaths = new Array();
|
||
let smallestPathLength = Number.MAX_SAFE_INTEGER;
|
||
// split each of the search paths using the platform specific separator
|
||
for (const searchPath of searchPaths) {
|
||
core_1.debug(`Using search path ${searchPath}`);
|
||
const splitSearchPath = path.normalize(searchPath).split(path.sep);
|
||
// keep track of the smallest path length so that we don't accidentally later go out of bounds
|
||
smallestPathLength = Math.min(smallestPathLength, splitSearchPath.length);
|
||
splitPaths.push(splitSearchPath);
|
||
}
|
||
// on Unix-like file systems, the file separator exists at the beginning of the file path, make sure to preserve it
|
||
if (searchPaths[0].startsWith(path.sep)) {
|
||
commonPaths.push(path.sep);
|
||
}
|
||
let splitIndex = 0;
|
||
// function to check if the paths are the same at a specific index
|
||
function isPathTheSame() {
|
||
const compare = splitPaths[0][splitIndex];
|
||
for (let i = 1; i < splitPaths.length; i++) {
|
||
if (compare !== splitPaths[i][splitIndex]) {
|
||
// a non-common index has been reached
|
||
return false;
|
||
}
|
||
}
|
||
return true;
|
||
}
|
||
// loop over all the search paths until there is a non-common ancestor or we go out of bounds
|
||
while (splitIndex < smallestPathLength) {
|
||
if (!isPathTheSame()) {
|
||
break;
|
||
}
|
||
// if all are the same, add to the end result & increment the index
|
||
commonPaths.push(splitPaths[0][splitIndex]);
|
||
splitIndex++;
|
||
}
|
||
return path.join(...commonPaths);
|
||
}
|
||
function findFilesToUpload(searchPath, globOptions) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
const searchResults = [];
|
||
const globber = yield glob.create(searchPath, globOptions || getDefaultGlobOptions());
|
||
const rawSearchResults = yield globber.glob();
|
||
/*
|
||
Files are saved with case insensitivity. Uploading both a.txt and A.txt will files to be overwritten
|
||
Detect any files that could be overwritten for user awareness
|
||
*/
|
||
const set = new Set();
|
||
/*
|
||
Directories will be rejected if attempted to be uploaded. This includes just empty
|
||
directories so filter any directories out from the raw search results
|
||
*/
|
||
for (const searchResult of rawSearchResults) {
|
||
const fileStats = yield stats(searchResult);
|
||
// isDirectory() returns false for symlinks if using fs.lstat(), make sure to use fs.stat() instead
|
||
if (!fileStats.isDirectory()) {
|
||
core_1.debug(`File:${searchResult} was found using the provided searchPath`);
|
||
searchResults.push(searchResult);
|
||
// detect any files that would be overwritten because of case insensitivity
|
||
if (set.has(searchResult.toLowerCase())) {
|
||
core_1.info(`Uploads are case insensitive: ${searchResult} was detected that it will be overwritten by another file with the same path`);
|
||
}
|
||
else {
|
||
set.add(searchResult.toLowerCase());
|
||
}
|
||
}
|
||
else {
|
||
core_1.debug(`Removing ${searchResult} from rawSearchResults because it is a directory`);
|
||
}
|
||
}
|
||
// Calculate the root directory for the artifact using the search paths that were utilized
|
||
const searchPaths = globber.getSearchPaths();
|
||
if (searchPaths.length > 1) {
|
||
core_1.info(`Multiple search paths detected. Calculating the least common ancestor of all paths`);
|
||
const lcaSearchPath = getMultiPathLCA(searchPaths);
|
||
core_1.info(`The least common ancestor is ${lcaSearchPath}. This will be the root directory of the artifact`);
|
||
return {
|
||
filesToUpload: searchResults,
|
||
rootDirectory: lcaSearchPath
|
||
};
|
||
}
|
||
/*
|
||
Special case for a single file artifact that is uploaded without a directory or wildcard pattern. The directory structure is
|
||
not preserved and the root directory will be the single files parent directory
|
||
*/
|
||
if (searchResults.length === 1 && searchPaths[0] === searchResults[0]) {
|
||
return {
|
||
filesToUpload: searchResults,
|
||
rootDirectory: path_1.dirname(searchResults[0])
|
||
};
|
||
}
|
||
return {
|
||
filesToUpload: searchResults,
|
||
rootDirectory: searchPaths[0]
|
||
};
|
||
});
|
||
}
|
||
exports.findFilesToUpload = findFilesToUpload;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 583:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __importStar = (this && this.__importStar) || function (mod) {
|
||
if (mod && mod.__esModule) return mod;
|
||
var result = {};
|
||
if (mod != null) for (var k in mod) if (Object.hasOwnProperty.call(mod, k)) result[k] = mod[k];
|
||
result["default"] = mod;
|
||
return result;
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const core = __importStar(__webpack_require__(470));
|
||
const constants_1 = __webpack_require__(694);
|
||
/**
|
||
* Helper to get all the inputs for the action
|
||
*/
|
||
function getInputs() {
|
||
const name = core.getInput(constants_1.Inputs.Name);
|
||
const path = core.getInput(constants_1.Inputs.Path, { required: true });
|
||
const token = core.getInput(constants_1.Inputs.Token);
|
||
const ifNoFilesFound = core.getInput(constants_1.Inputs.IfNoFilesFound);
|
||
const noFileBehavior = constants_1.NoFileOptions[ifNoFilesFound];
|
||
if (!noFileBehavior) {
|
||
core.setFailed(`Unrecognized ${constants_1.Inputs.IfNoFilesFound} input. Provided: ${ifNoFilesFound}. Available options: ${Object.keys(constants_1.NoFileOptions)}`);
|
||
}
|
||
const inputs = {
|
||
artifactName: name,
|
||
searchPath: path,
|
||
ifNoFilesFound: noFileBehavior,
|
||
token: token
|
||
};
|
||
const retentionDaysStr = core.getInput(constants_1.Inputs.RetentionDays);
|
||
if (retentionDaysStr) {
|
||
inputs.retentionDays = parseInt(retentionDaysStr);
|
||
if (isNaN(inputs.retentionDays)) {
|
||
core.setFailed('Invalid retention-days');
|
||
}
|
||
}
|
||
return inputs;
|
||
}
|
||
exports.getInputs = getInputs;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 589:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var utils = __webpack_require__(35);
|
||
var defaults = __webpack_require__(529);
|
||
|
||
/**
|
||
* Transform the data for a request or a response
|
||
*
|
||
* @param {Object|String} data The data to be transformed
|
||
* @param {Array} headers The headers for the request or response
|
||
* @param {Array|Function} fns A single function or Array of functions
|
||
* @returns {*} The resulting transformed data
|
||
*/
|
||
module.exports = function transformData(data, headers, fns) {
|
||
var context = this || defaults;
|
||
/*eslint no-param-reassign:0*/
|
||
utils.forEach(fns, function transform(fn) {
|
||
data = fn.call(context, data, headers);
|
||
});
|
||
|
||
return data;
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 590:
|
||
/***/ (function(module) {
|
||
|
||
"use strict";
|
||
|
||
|
||
/**
|
||
* Determines whether the specified URL is absolute
|
||
*
|
||
* @param {string} url The URL to test
|
||
* @returns {boolean} True if the specified URL is absolute, otherwise false
|
||
*/
|
||
module.exports = function isAbsoluteURL(url) {
|
||
// A URL is considered absolute if it begins with "<scheme>://" or "//" (protocol-relative URL).
|
||
// RFC 3986 defines scheme name as a sequence of characters beginning with a letter and followed
|
||
// by any combination of letters, digits, plus, period, or hyphen.
|
||
return /^([a-z][a-z\d\+\-\.]*:)?\/\//i.test(url);
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 597:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const pathHelper = __webpack_require__(972);
|
||
const internal_match_kind_1 = __webpack_require__(327);
|
||
const IS_WINDOWS = process.platform === 'win32';
|
||
/**
|
||
* Given an array of patterns, returns an array of paths to search.
|
||
* Duplicates and paths under other included paths are filtered out.
|
||
*/
|
||
function getSearchPaths(patterns) {
|
||
// Ignore negate patterns
|
||
patterns = patterns.filter(x => !x.negate);
|
||
// Create a map of all search paths
|
||
const searchPathMap = {};
|
||
for (const pattern of patterns) {
|
||
const key = IS_WINDOWS
|
||
? pattern.searchPath.toUpperCase()
|
||
: pattern.searchPath;
|
||
searchPathMap[key] = 'candidate';
|
||
}
|
||
const result = [];
|
||
for (const pattern of patterns) {
|
||
// Check if already included
|
||
const key = IS_WINDOWS
|
||
? pattern.searchPath.toUpperCase()
|
||
: pattern.searchPath;
|
||
if (searchPathMap[key] === 'included') {
|
||
continue;
|
||
}
|
||
// Check for an ancestor search path
|
||
let foundAncestor = false;
|
||
let tempKey = key;
|
||
let parent = pathHelper.dirname(tempKey);
|
||
while (parent !== tempKey) {
|
||
if (searchPathMap[parent]) {
|
||
foundAncestor = true;
|
||
break;
|
||
}
|
||
tempKey = parent;
|
||
parent = pathHelper.dirname(tempKey);
|
||
}
|
||
// Include the search pattern in the result
|
||
if (!foundAncestor) {
|
||
result.push(pattern.searchPath);
|
||
searchPathMap[key] = 'included';
|
||
}
|
||
}
|
||
return result;
|
||
}
|
||
exports.getSearchPaths = getSearchPaths;
|
||
/**
|
||
* Matches the patterns against the path
|
||
*/
|
||
function match(patterns, itemPath) {
|
||
let result = internal_match_kind_1.MatchKind.None;
|
||
for (const pattern of patterns) {
|
||
if (pattern.negate) {
|
||
result &= ~pattern.match(itemPath);
|
||
}
|
||
else {
|
||
result |= pattern.match(itemPath);
|
||
}
|
||
}
|
||
return result;
|
||
}
|
||
exports.match = match;
|
||
/**
|
||
* Checks whether to descend further into the directory
|
||
*/
|
||
function partialMatch(patterns, itemPath) {
|
||
return patterns.some(x => !x.negate && x.partialMatch(itemPath));
|
||
}
|
||
exports.partialMatch = partialMatch;
|
||
//# sourceMappingURL=internal-pattern-helper.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 601:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const core = __webpack_require__(470);
|
||
/**
|
||
* Returns a copy with defaults filled in.
|
||
*/
|
||
function getOptions(copy) {
|
||
const result = {
|
||
followSymbolicLinks: true,
|
||
implicitDescendants: true,
|
||
omitBrokenSymbolicLinks: true
|
||
};
|
||
if (copy) {
|
||
if (typeof copy.followSymbolicLinks === 'boolean') {
|
||
result.followSymbolicLinks = copy.followSymbolicLinks;
|
||
core.debug(`followSymbolicLinks '${result.followSymbolicLinks}'`);
|
||
}
|
||
if (typeof copy.implicitDescendants === 'boolean') {
|
||
result.implicitDescendants = copy.implicitDescendants;
|
||
core.debug(`implicitDescendants '${result.implicitDescendants}'`);
|
||
}
|
||
if (typeof copy.omitBrokenSymbolicLinks === 'boolean') {
|
||
result.omitBrokenSymbolicLinks = copy.omitBrokenSymbolicLinks;
|
||
core.debug(`omitBrokenSymbolicLinks '${result.omitBrokenSymbolicLinks}'`);
|
||
}
|
||
}
|
||
return result;
|
||
}
|
||
exports.getOptions = getOptions;
|
||
//# sourceMappingURL=internal-glob-options-helper.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 605:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("http");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 608:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
||
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
||
return new (P || (P = Promise))(function (resolve, reject) {
|
||
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
||
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
||
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
||
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
||
});
|
||
};
|
||
var __importStar = (this && this.__importStar) || function (mod) {
|
||
if (mod && mod.__esModule) return mod;
|
||
var result = {};
|
||
if (mod != null) for (var k in mod) if (Object.hasOwnProperty.call(mod, k)) result[k] = mod[k];
|
||
result["default"] = mod;
|
||
return result;
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const fs = __importStar(__webpack_require__(747));
|
||
const core = __importStar(__webpack_require__(470));
|
||
const tmp = __importStar(__webpack_require__(875));
|
||
const stream = __importStar(__webpack_require__(413));
|
||
const utils_1 = __webpack_require__(870);
|
||
const config_variables_1 = __webpack_require__(401);
|
||
const util_1 = __webpack_require__(669);
|
||
const url_1 = __webpack_require__(835);
|
||
const perf_hooks_1 = __webpack_require__(630);
|
||
const status_reporter_1 = __webpack_require__(176);
|
||
const http_client_1 = __webpack_require__(539);
|
||
const http_manager_1 = __webpack_require__(452);
|
||
const upload_gzip_1 = __webpack_require__(647);
|
||
const requestUtils_1 = __webpack_require__(489);
|
||
const stat = util_1.promisify(fs.stat);
|
||
class UploadHttpClient {
|
||
constructor() {
|
||
this.uploadHttpManager = new http_manager_1.HttpManager(config_variables_1.getUploadFileConcurrency(), '@actions/artifact-upload');
|
||
this.statusReporter = new status_reporter_1.StatusReporter(10000);
|
||
}
|
||
/**
|
||
* Creates a file container for the new artifact in the remote blob storage/file service
|
||
* @param {string} artifactName Name of the artifact being created
|
||
* @returns The response from the Artifact Service if the file container was successfully created
|
||
*/
|
||
createArtifactInFileContainer(artifactName, options) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
const parameters = {
|
||
Type: 'actions_storage',
|
||
Name: artifactName
|
||
};
|
||
// calculate retention period
|
||
if (options && options.retentionDays) {
|
||
const maxRetentionStr = config_variables_1.getRetentionDays();
|
||
parameters.RetentionDays = utils_1.getProperRetention(options.retentionDays, maxRetentionStr);
|
||
}
|
||
const data = JSON.stringify(parameters, null, 2);
|
||
const artifactUrl = utils_1.getArtifactUrl();
|
||
// use the first client from the httpManager, `keep-alive` is not used so the connection will close immediately
|
||
const client = this.uploadHttpManager.getClient(0);
|
||
const headers = utils_1.getUploadHeaders('application/json', false);
|
||
// Extra information to display when a particular HTTP code is returned
|
||
// If a 403 is returned when trying to create a file container, the customer has exceeded
|
||
// their storage quota so no new artifact containers can be created
|
||
const customErrorMessages = new Map([
|
||
[
|
||
http_client_1.HttpCodes.Forbidden,
|
||
'Artifact storage quota has been hit. Unable to upload any new artifacts'
|
||
],
|
||
[
|
||
http_client_1.HttpCodes.BadRequest,
|
||
`The artifact name ${artifactName} is not valid. Request URL ${artifactUrl}`
|
||
]
|
||
]);
|
||
const response = yield requestUtils_1.retryHttpClientRequest('Create Artifact Container', () => __awaiter(this, void 0, void 0, function* () { return client.post(artifactUrl, data, headers); }), customErrorMessages);
|
||
const body = yield response.readBody();
|
||
return JSON.parse(body);
|
||
});
|
||
}
|
||
/**
|
||
* Concurrently upload all of the files in chunks
|
||
* @param {string} uploadUrl Base Url for the artifact that was created
|
||
* @param {SearchResult[]} filesToUpload A list of information about the files being uploaded
|
||
* @returns The size of all the files uploaded in bytes
|
||
*/
|
||
uploadArtifactToFileContainer(uploadUrl, filesToUpload, options) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
const FILE_CONCURRENCY = config_variables_1.getUploadFileConcurrency();
|
||
const MAX_CHUNK_SIZE = config_variables_1.getUploadChunkSize();
|
||
core.debug(`File Concurrency: ${FILE_CONCURRENCY}, and Chunk Size: ${MAX_CHUNK_SIZE}`);
|
||
const parameters = [];
|
||
// by default, file uploads will continue if there is an error unless specified differently in the options
|
||
let continueOnError = true;
|
||
if (options) {
|
||
if (options.continueOnError === false) {
|
||
continueOnError = false;
|
||
}
|
||
}
|
||
// prepare the necessary parameters to upload all the files
|
||
for (const file of filesToUpload) {
|
||
const resourceUrl = new url_1.URL(uploadUrl);
|
||
resourceUrl.searchParams.append('itemPath', file.uploadFilePath);
|
||
parameters.push({
|
||
file: file.absoluteFilePath,
|
||
resourceUrl: resourceUrl.toString(),
|
||
maxChunkSize: MAX_CHUNK_SIZE,
|
||
continueOnError
|
||
});
|
||
}
|
||
const parallelUploads = [...new Array(FILE_CONCURRENCY).keys()];
|
||
const failedItemsToReport = [];
|
||
let currentFile = 0;
|
||
let completedFiles = 0;
|
||
let uploadFileSize = 0;
|
||
let totalFileSize = 0;
|
||
let abortPendingFileUploads = false;
|
||
this.statusReporter.setTotalNumberOfFilesToProcess(filesToUpload.length);
|
||
this.statusReporter.start();
|
||
// only allow a certain amount of files to be uploaded at once, this is done to reduce potential errors
|
||
yield Promise.all(parallelUploads.map((index) => __awaiter(this, void 0, void 0, function* () {
|
||
while (currentFile < filesToUpload.length) {
|
||
const currentFileParameters = parameters[currentFile];
|
||
currentFile += 1;
|
||
if (abortPendingFileUploads) {
|
||
failedItemsToReport.push(currentFileParameters.file);
|
||
continue;
|
||
}
|
||
const startTime = perf_hooks_1.performance.now();
|
||
const uploadFileResult = yield this.uploadFileAsync(index, currentFileParameters);
|
||
if (core.isDebug()) {
|
||
core.debug(`File: ${++completedFiles}/${filesToUpload.length}. ${currentFileParameters.file} took ${(perf_hooks_1.performance.now() - startTime).toFixed(3)} milliseconds to finish upload`);
|
||
}
|
||
uploadFileSize += uploadFileResult.successfulUploadSize;
|
||
totalFileSize += uploadFileResult.totalSize;
|
||
if (uploadFileResult.isSuccess === false) {
|
||
failedItemsToReport.push(currentFileParameters.file);
|
||
if (!continueOnError) {
|
||
// fail fast
|
||
core.error(`aborting artifact upload`);
|
||
abortPendingFileUploads = true;
|
||
}
|
||
}
|
||
this.statusReporter.incrementProcessedCount();
|
||
}
|
||
})));
|
||
this.statusReporter.stop();
|
||
// done uploading, safety dispose all connections
|
||
this.uploadHttpManager.disposeAndReplaceAllClients();
|
||
core.info(`Total size of all the files uploaded is ${uploadFileSize} bytes`);
|
||
return {
|
||
uploadSize: uploadFileSize,
|
||
totalSize: totalFileSize,
|
||
failedItems: failedItemsToReport
|
||
};
|
||
});
|
||
}
|
||
/**
|
||
* Asynchronously uploads a file. The file is compressed and uploaded using GZip if it is determined to save space.
|
||
* If the upload file is bigger than the max chunk size it will be uploaded via multiple calls
|
||
* @param {number} httpClientIndex The index of the httpClient that is being used to make all of the calls
|
||
* @param {UploadFileParameters} parameters Information about the file that needs to be uploaded
|
||
* @returns The size of the file that was uploaded in bytes along with any failed uploads
|
||
*/
|
||
uploadFileAsync(httpClientIndex, parameters) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
const totalFileSize = (yield stat(parameters.file)).size;
|
||
let offset = 0;
|
||
let isUploadSuccessful = true;
|
||
let failedChunkSizes = 0;
|
||
let uploadFileSize = 0;
|
||
let isGzip = true;
|
||
// the file that is being uploaded is less than 64k in size, to increase throughput and to minimize disk I/O
|
||
// for creating a new GZip file, an in-memory buffer is used for compression
|
||
if (totalFileSize < 65536) {
|
||
const buffer = yield upload_gzip_1.createGZipFileInBuffer(parameters.file);
|
||
//An open stream is needed in the event of a failure and we need to retry. If a NodeJS.ReadableStream is directly passed in,
|
||
// it will not properly get reset to the start of the stream if a chunk upload needs to be retried
|
||
let openUploadStream;
|
||
if (totalFileSize < buffer.byteLength) {
|
||
// compression did not help with reducing the size, use a readable stream from the original file for upload
|
||
openUploadStream = () => fs.createReadStream(parameters.file);
|
||
isGzip = false;
|
||
uploadFileSize = totalFileSize;
|
||
}
|
||
else {
|
||
// create a readable stream using a PassThrough stream that is both readable and writable
|
||
openUploadStream = () => {
|
||
const passThrough = new stream.PassThrough();
|
||
passThrough.end(buffer);
|
||
return passThrough;
|
||
};
|
||
uploadFileSize = buffer.byteLength;
|
||
}
|
||
const result = yield this.uploadChunk(httpClientIndex, parameters.resourceUrl, openUploadStream, 0, uploadFileSize - 1, uploadFileSize, isGzip, totalFileSize);
|
||
if (!result) {
|
||
// chunk failed to upload
|
||
isUploadSuccessful = false;
|
||
failedChunkSizes += uploadFileSize;
|
||
core.warning(`Aborting upload for ${parameters.file} due to failure`);
|
||
}
|
||
return {
|
||
isSuccess: isUploadSuccessful,
|
||
successfulUploadSize: uploadFileSize - failedChunkSizes,
|
||
totalSize: totalFileSize
|
||
};
|
||
}
|
||
else {
|
||
// the file that is being uploaded is greater than 64k in size, a temporary file gets created on disk using the
|
||
// npm tmp-promise package and this file gets used to create a GZipped file
|
||
const tempFile = yield tmp.file();
|
||
// create a GZip file of the original file being uploaded, the original file should not be modified in any way
|
||
uploadFileSize = yield upload_gzip_1.createGZipFileOnDisk(parameters.file, tempFile.path);
|
||
let uploadFilePath = tempFile.path;
|
||
// compression did not help with size reduction, use the original file for upload and delete the temp GZip file
|
||
if (totalFileSize < uploadFileSize) {
|
||
uploadFileSize = totalFileSize;
|
||
uploadFilePath = parameters.file;
|
||
isGzip = false;
|
||
}
|
||
let abortFileUpload = false;
|
||
// upload only a single chunk at a time
|
||
while (offset < uploadFileSize) {
|
||
const chunkSize = Math.min(uploadFileSize - offset, parameters.maxChunkSize);
|
||
// if an individual file is greater than 100MB (1024*1024*100) in size, display extra information about the upload status
|
||
if (uploadFileSize > 104857600) {
|
||
this.statusReporter.updateLargeFileStatus(parameters.file, offset, uploadFileSize);
|
||
}
|
||
const start = offset;
|
||
const end = offset + chunkSize - 1;
|
||
offset += parameters.maxChunkSize;
|
||
if (abortFileUpload) {
|
||
// if we don't want to continue in the event of an error, any pending upload chunks will be marked as failed
|
||
failedChunkSizes += chunkSize;
|
||
continue;
|
||
}
|
||
const result = yield this.uploadChunk(httpClientIndex, parameters.resourceUrl, () => fs.createReadStream(uploadFilePath, {
|
||
start,
|
||
end,
|
||
autoClose: false
|
||
}), start, end, uploadFileSize, isGzip, totalFileSize);
|
||
if (!result) {
|
||
// Chunk failed to upload, report as failed and do not continue uploading any more chunks for the file. It is possible that part of a chunk was
|
||
// successfully uploaded so the server may report a different size for what was uploaded
|
||
isUploadSuccessful = false;
|
||
failedChunkSizes += chunkSize;
|
||
core.warning(`Aborting upload for ${parameters.file} due to failure`);
|
||
abortFileUpload = true;
|
||
}
|
||
}
|
||
// Delete the temporary file that was created as part of the upload. If the temp file does not get manually deleted by
|
||
// calling cleanup, it gets removed when the node process exits. For more info see: https://www.npmjs.com/package/tmp-promise#about
|
||
yield tempFile.cleanup();
|
||
return {
|
||
isSuccess: isUploadSuccessful,
|
||
successfulUploadSize: uploadFileSize - failedChunkSizes,
|
||
totalSize: totalFileSize
|
||
};
|
||
}
|
||
});
|
||
}
|
||
/**
|
||
* Uploads a chunk of an individual file to the specified resourceUrl. If the upload fails and the status code
|
||
* indicates a retryable status, we try to upload the chunk as well
|
||
* @param {number} httpClientIndex The index of the httpClient being used to make all the necessary calls
|
||
* @param {string} resourceUrl Url of the resource that the chunk will be uploaded to
|
||
* @param {NodeJS.ReadableStream} openStream Stream of the file that will be uploaded
|
||
* @param {number} start Starting byte index of file that the chunk belongs to
|
||
* @param {number} end Ending byte index of file that the chunk belongs to
|
||
* @param {number} uploadFileSize Total size of the file in bytes that is being uploaded
|
||
* @param {boolean} isGzip Denotes if we are uploading a Gzip compressed stream
|
||
* @param {number} totalFileSize Original total size of the file that is being uploaded
|
||
* @returns if the chunk was successfully uploaded
|
||
*/
|
||
uploadChunk(httpClientIndex, resourceUrl, openStream, start, end, uploadFileSize, isGzip, totalFileSize) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
// prepare all the necessary headers before making any http call
|
||
const headers = utils_1.getUploadHeaders('application/octet-stream', true, isGzip, totalFileSize, end - start + 1, utils_1.getContentRange(start, end, uploadFileSize));
|
||
const uploadChunkRequest = () => __awaiter(this, void 0, void 0, function* () {
|
||
const client = this.uploadHttpManager.getClient(httpClientIndex);
|
||
return yield client.sendStream('PUT', resourceUrl, openStream(), headers);
|
||
});
|
||
let retryCount = 0;
|
||
const retryLimit = config_variables_1.getRetryLimit();
|
||
// Increments the current retry count and then checks if the retry limit has been reached
|
||
// If there have been too many retries, fail so the download stops
|
||
const incrementAndCheckRetryLimit = (response) => {
|
||
retryCount++;
|
||
if (retryCount > retryLimit) {
|
||
if (response) {
|
||
utils_1.displayHttpDiagnostics(response);
|
||
}
|
||
core.info(`Retry limit has been reached for chunk at offset ${start} to ${resourceUrl}`);
|
||
return true;
|
||
}
|
||
return false;
|
||
};
|
||
const backOff = (retryAfterValue) => __awaiter(this, void 0, void 0, function* () {
|
||
this.uploadHttpManager.disposeAndReplaceClient(httpClientIndex);
|
||
if (retryAfterValue) {
|
||
core.info(`Backoff due to too many requests, retry #${retryCount}. Waiting for ${retryAfterValue} milliseconds before continuing the upload`);
|
||
yield utils_1.sleep(retryAfterValue);
|
||
}
|
||
else {
|
||
const backoffTime = utils_1.getExponentialRetryTimeInMilliseconds(retryCount);
|
||
core.info(`Exponential backoff for retry #${retryCount}. Waiting for ${backoffTime} milliseconds before continuing the upload at offset ${start}`);
|
||
yield utils_1.sleep(backoffTime);
|
||
}
|
||
core.info(`Finished backoff for retry #${retryCount}, continuing with upload`);
|
||
return;
|
||
});
|
||
// allow for failed chunks to be retried multiple times
|
||
while (retryCount <= retryLimit) {
|
||
let response;
|
||
try {
|
||
response = yield uploadChunkRequest();
|
||
}
|
||
catch (error) {
|
||
// if an error is caught, it is usually indicative of a timeout so retry the upload
|
||
core.info(`An error has been caught http-client index ${httpClientIndex}, retrying the upload`);
|
||
// eslint-disable-next-line no-console
|
||
console.log(error);
|
||
if (incrementAndCheckRetryLimit()) {
|
||
return false;
|
||
}
|
||
yield backOff();
|
||
continue;
|
||
}
|
||
// Always read the body of the response. There is potential for a resource leak if the body is not read which will
|
||
// result in the connection remaining open along with unintended consequences when trying to dispose of the client
|
||
yield response.readBody();
|
||
if (utils_1.isSuccessStatusCode(response.message.statusCode)) {
|
||
return true;
|
||
}
|
||
else if (utils_1.isRetryableStatusCode(response.message.statusCode)) {
|
||
core.info(`A ${response.message.statusCode} status code has been received, will attempt to retry the upload`);
|
||
if (incrementAndCheckRetryLimit(response)) {
|
||
return false;
|
||
}
|
||
utils_1.isThrottledStatusCode(response.message.statusCode)
|
||
? yield backOff(utils_1.tryGetRetryAfterValueTimeInMilliseconds(response.message.headers))
|
||
: yield backOff();
|
||
}
|
||
else {
|
||
core.error(`Unexpected response. Unable to upload chunk to ${resourceUrl}`);
|
||
utils_1.displayHttpDiagnostics(response);
|
||
return false;
|
||
}
|
||
}
|
||
return false;
|
||
});
|
||
}
|
||
/**
|
||
* Updates the size of the artifact from -1 which was initially set when the container was first created for the artifact.
|
||
* Updating the size indicates that we are done uploading all the contents of the artifact
|
||
*/
|
||
patchArtifactSize(size, artifactName) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
const resourceUrl = new url_1.URL(utils_1.getArtifactUrl());
|
||
resourceUrl.searchParams.append('artifactName', artifactName);
|
||
const parameters = { Size: size };
|
||
const data = JSON.stringify(parameters, null, 2);
|
||
core.debug(`URL is ${resourceUrl.toString()}`);
|
||
// use the first client from the httpManager, `keep-alive` is not used so the connection will close immediately
|
||
const client = this.uploadHttpManager.getClient(0);
|
||
const headers = utils_1.getUploadHeaders('application/json', false);
|
||
// Extra information to display when a particular HTTP code is returned
|
||
const customErrorMessages = new Map([
|
||
[
|
||
http_client_1.HttpCodes.NotFound,
|
||
`An Artifact with the name ${artifactName} was not found`
|
||
]
|
||
]);
|
||
// TODO retry for all possible response codes, the artifact upload is pretty much complete so it at all costs we should try to finish this
|
||
const response = yield requestUtils_1.retryHttpClientRequest('Finalize artifact upload', () => __awaiter(this, void 0, void 0, function* () { return client.patch(resourceUrl.toString(), data, headers); }), customErrorMessages);
|
||
yield response.readBody();
|
||
core.debug(`Artifact ${artifactName} has been successfully uploaded, total size in bytes: ${size}`);
|
||
});
|
||
}
|
||
}
|
||
exports.UploadHttpClient = UploadHttpClient;
|
||
//# sourceMappingURL=upload-http-client.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 614:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("events");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 621:
|
||
/***/ (function(module) {
|
||
|
||
"use strict";
|
||
|
||
module.exports = balanced;
|
||
function balanced(a, b, str) {
|
||
if (a instanceof RegExp) a = maybeMatch(a, str);
|
||
if (b instanceof RegExp) b = maybeMatch(b, str);
|
||
|
||
var r = range(a, b, str);
|
||
|
||
return r && {
|
||
start: r[0],
|
||
end: r[1],
|
||
pre: str.slice(0, r[0]),
|
||
body: str.slice(r[0] + a.length, r[1]),
|
||
post: str.slice(r[1] + b.length)
|
||
};
|
||
}
|
||
|
||
function maybeMatch(reg, str) {
|
||
var m = str.match(reg);
|
||
return m ? m[0] : null;
|
||
}
|
||
|
||
balanced.range = range;
|
||
function range(a, b, str) {
|
||
var begs, beg, left, right, result;
|
||
var ai = str.indexOf(a);
|
||
var bi = str.indexOf(b, ai + 1);
|
||
var i = ai;
|
||
|
||
if (ai >= 0 && bi > 0) {
|
||
begs = [];
|
||
left = str.length;
|
||
|
||
while (i >= 0 && !result) {
|
||
if (i == ai) {
|
||
begs.push(i);
|
||
ai = str.indexOf(a, i + 1);
|
||
} else if (begs.length == 1) {
|
||
result = [ begs.pop(), bi ];
|
||
} else {
|
||
beg = begs.pop();
|
||
if (beg < left) {
|
||
left = beg;
|
||
right = bi;
|
||
}
|
||
|
||
bi = str.indexOf(b, i + 1);
|
||
}
|
||
|
||
i = ai < bi && ai >= 0 ? ai : bi;
|
||
}
|
||
|
||
if (begs.length) {
|
||
result = [ left, right ];
|
||
}
|
||
}
|
||
|
||
return result;
|
||
}
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 622:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("path");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 630:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("perf_hooks");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 631:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var utils = __webpack_require__(35);
|
||
|
||
// Headers whose duplicates are ignored by node
|
||
// c.f. https://nodejs.org/api/http.html#http_message_headers
|
||
var ignoreDuplicateOf = [
|
||
'age', 'authorization', 'content-length', 'content-type', 'etag',
|
||
'expires', 'from', 'host', 'if-modified-since', 'if-unmodified-since',
|
||
'last-modified', 'location', 'max-forwards', 'proxy-authorization',
|
||
'referer', 'retry-after', 'user-agent'
|
||
];
|
||
|
||
/**
|
||
* Parse headers into an object
|
||
*
|
||
* ```
|
||
* Date: Wed, 27 Aug 2014 08:58:49 GMT
|
||
* Content-Type: application/json
|
||
* Connection: keep-alive
|
||
* Transfer-Encoding: chunked
|
||
* ```
|
||
*
|
||
* @param {String} headers Headers needing to be parsed
|
||
* @returns {Object} Headers parsed into an object
|
||
*/
|
||
module.exports = function parseHeaders(headers) {
|
||
var parsed = {};
|
||
var key;
|
||
var val;
|
||
var i;
|
||
|
||
if (!headers) { return parsed; }
|
||
|
||
utils.forEach(headers.split('\n'), function parser(line) {
|
||
i = line.indexOf(':');
|
||
key = utils.trim(line.substr(0, i)).toLowerCase();
|
||
val = utils.trim(line.substr(i + 1));
|
||
|
||
if (key) {
|
||
if (parsed[key] && ignoreDuplicateOf.indexOf(key) >= 0) {
|
||
return;
|
||
}
|
||
if (key === 'set-cookie') {
|
||
parsed[key] = (parsed[key] ? parsed[key] : []).concat([val]);
|
||
} else {
|
||
parsed[key] = parsed[key] ? parsed[key] + ', ' + val : val;
|
||
}
|
||
}
|
||
});
|
||
|
||
return parsed;
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 644:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
exports.alphasort = alphasort
|
||
exports.alphasorti = alphasorti
|
||
exports.setopts = setopts
|
||
exports.ownProp = ownProp
|
||
exports.makeAbs = makeAbs
|
||
exports.finish = finish
|
||
exports.mark = mark
|
||
exports.isIgnored = isIgnored
|
||
exports.childrenIgnored = childrenIgnored
|
||
|
||
function ownProp (obj, field) {
|
||
return Object.prototype.hasOwnProperty.call(obj, field)
|
||
}
|
||
|
||
var path = __webpack_require__(622)
|
||
var minimatch = __webpack_require__(93)
|
||
var isAbsolute = __webpack_require__(681)
|
||
var Minimatch = minimatch.Minimatch
|
||
|
||
function alphasorti (a, b) {
|
||
return a.toLowerCase().localeCompare(b.toLowerCase())
|
||
}
|
||
|
||
function alphasort (a, b) {
|
||
return a.localeCompare(b)
|
||
}
|
||
|
||
function setupIgnores (self, options) {
|
||
self.ignore = options.ignore || []
|
||
|
||
if (!Array.isArray(self.ignore))
|
||
self.ignore = [self.ignore]
|
||
|
||
if (self.ignore.length) {
|
||
self.ignore = self.ignore.map(ignoreMap)
|
||
}
|
||
}
|
||
|
||
// ignore patterns are always in dot:true mode.
|
||
function ignoreMap (pattern) {
|
||
var gmatcher = null
|
||
if (pattern.slice(-3) === '/**') {
|
||
var gpattern = pattern.replace(/(\/\*\*)+$/, '')
|
||
gmatcher = new Minimatch(gpattern, { dot: true })
|
||
}
|
||
|
||
return {
|
||
matcher: new Minimatch(pattern, { dot: true }),
|
||
gmatcher: gmatcher
|
||
}
|
||
}
|
||
|
||
function setopts (self, pattern, options) {
|
||
if (!options)
|
||
options = {}
|
||
|
||
// base-matching: just use globstar for that.
|
||
if (options.matchBase && -1 === pattern.indexOf("/")) {
|
||
if (options.noglobstar) {
|
||
throw new Error("base matching requires globstar")
|
||
}
|
||
pattern = "**/" + pattern
|
||
}
|
||
|
||
self.silent = !!options.silent
|
||
self.pattern = pattern
|
||
self.strict = options.strict !== false
|
||
self.realpath = !!options.realpath
|
||
self.realpathCache = options.realpathCache || Object.create(null)
|
||
self.follow = !!options.follow
|
||
self.dot = !!options.dot
|
||
self.mark = !!options.mark
|
||
self.nodir = !!options.nodir
|
||
if (self.nodir)
|
||
self.mark = true
|
||
self.sync = !!options.sync
|
||
self.nounique = !!options.nounique
|
||
self.nonull = !!options.nonull
|
||
self.nosort = !!options.nosort
|
||
self.nocase = !!options.nocase
|
||
self.stat = !!options.stat
|
||
self.noprocess = !!options.noprocess
|
||
self.absolute = !!options.absolute
|
||
|
||
self.maxLength = options.maxLength || Infinity
|
||
self.cache = options.cache || Object.create(null)
|
||
self.statCache = options.statCache || Object.create(null)
|
||
self.symlinks = options.symlinks || Object.create(null)
|
||
|
||
setupIgnores(self, options)
|
||
|
||
self.changedCwd = false
|
||
var cwd = process.cwd()
|
||
if (!ownProp(options, "cwd"))
|
||
self.cwd = cwd
|
||
else {
|
||
self.cwd = path.resolve(options.cwd)
|
||
self.changedCwd = self.cwd !== cwd
|
||
}
|
||
|
||
self.root = options.root || path.resolve(self.cwd, "/")
|
||
self.root = path.resolve(self.root)
|
||
if (process.platform === "win32")
|
||
self.root = self.root.replace(/\\/g, "/")
|
||
|
||
// TODO: is an absolute `cwd` supposed to be resolved against `root`?
|
||
// e.g. { cwd: '/test', root: __dirname } === path.join(__dirname, '/test')
|
||
self.cwdAbs = isAbsolute(self.cwd) ? self.cwd : makeAbs(self, self.cwd)
|
||
if (process.platform === "win32")
|
||
self.cwdAbs = self.cwdAbs.replace(/\\/g, "/")
|
||
self.nomount = !!options.nomount
|
||
|
||
// disable comments and negation in Minimatch.
|
||
// Note that they are not supported in Glob itself anyway.
|
||
options.nonegate = true
|
||
options.nocomment = true
|
||
|
||
self.minimatch = new Minimatch(pattern, options)
|
||
self.options = self.minimatch.options
|
||
}
|
||
|
||
function finish (self) {
|
||
var nou = self.nounique
|
||
var all = nou ? [] : Object.create(null)
|
||
|
||
for (var i = 0, l = self.matches.length; i < l; i ++) {
|
||
var matches = self.matches[i]
|
||
if (!matches || Object.keys(matches).length === 0) {
|
||
if (self.nonull) {
|
||
// do like the shell, and spit out the literal glob
|
||
var literal = self.minimatch.globSet[i]
|
||
if (nou)
|
||
all.push(literal)
|
||
else
|
||
all[literal] = true
|
||
}
|
||
} else {
|
||
// had matches
|
||
var m = Object.keys(matches)
|
||
if (nou)
|
||
all.push.apply(all, m)
|
||
else
|
||
m.forEach(function (m) {
|
||
all[m] = true
|
||
})
|
||
}
|
||
}
|
||
|
||
if (!nou)
|
||
all = Object.keys(all)
|
||
|
||
if (!self.nosort)
|
||
all = all.sort(self.nocase ? alphasorti : alphasort)
|
||
|
||
// at *some* point we statted all of these
|
||
if (self.mark) {
|
||
for (var i = 0; i < all.length; i++) {
|
||
all[i] = self._mark(all[i])
|
||
}
|
||
if (self.nodir) {
|
||
all = all.filter(function (e) {
|
||
var notDir = !(/\/$/.test(e))
|
||
var c = self.cache[e] || self.cache[makeAbs(self, e)]
|
||
if (notDir && c)
|
||
notDir = c !== 'DIR' && !Array.isArray(c)
|
||
return notDir
|
||
})
|
||
}
|
||
}
|
||
|
||
if (self.ignore.length)
|
||
all = all.filter(function(m) {
|
||
return !isIgnored(self, m)
|
||
})
|
||
|
||
self.found = all
|
||
}
|
||
|
||
function mark (self, p) {
|
||
var abs = makeAbs(self, p)
|
||
var c = self.cache[abs]
|
||
var m = p
|
||
if (c) {
|
||
var isDir = c === 'DIR' || Array.isArray(c)
|
||
var slash = p.slice(-1) === '/'
|
||
|
||
if (isDir && !slash)
|
||
m += '/'
|
||
else if (!isDir && slash)
|
||
m = m.slice(0, -1)
|
||
|
||
if (m !== p) {
|
||
var mabs = makeAbs(self, m)
|
||
self.statCache[mabs] = self.statCache[abs]
|
||
self.cache[mabs] = self.cache[abs]
|
||
}
|
||
}
|
||
|
||
return m
|
||
}
|
||
|
||
// lotta situps...
|
||
function makeAbs (self, f) {
|
||
var abs = f
|
||
if (f.charAt(0) === '/') {
|
||
abs = path.join(self.root, f)
|
||
} else if (isAbsolute(f) || f === '') {
|
||
abs = f
|
||
} else if (self.changedCwd) {
|
||
abs = path.resolve(self.cwd, f)
|
||
} else {
|
||
abs = path.resolve(f)
|
||
}
|
||
|
||
if (process.platform === 'win32')
|
||
abs = abs.replace(/\\/g, '/')
|
||
|
||
return abs
|
||
}
|
||
|
||
|
||
// Return true, if pattern ends with globstar '**', for the accompanying parent directory.
|
||
// Ex:- If node_modules/** is the pattern, add 'node_modules' to ignore list along with it's contents
|
||
function isIgnored (self, path) {
|
||
if (!self.ignore.length)
|
||
return false
|
||
|
||
return self.ignore.some(function(item) {
|
||
return item.matcher.match(path) || !!(item.gmatcher && item.gmatcher.match(path))
|
||
})
|
||
}
|
||
|
||
function childrenIgnored (self, path) {
|
||
if (!self.ignore.length)
|
||
return false
|
||
|
||
return self.ignore.some(function(item) {
|
||
return !!(item.gmatcher && item.gmatcher.match(path))
|
||
})
|
||
}
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 647:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
||
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
||
return new (P || (P = Promise))(function (resolve, reject) {
|
||
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
||
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
||
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
||
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
||
});
|
||
};
|
||
var __asyncValues = (this && this.__asyncValues) || function (o) {
|
||
if (!Symbol.asyncIterator) throw new TypeError("Symbol.asyncIterator is not defined.");
|
||
var m = o[Symbol.asyncIterator], i;
|
||
return m ? m.call(o) : (o = typeof __values === "function" ? __values(o) : o[Symbol.iterator](), i = {}, verb("next"), verb("throw"), verb("return"), i[Symbol.asyncIterator] = function () { return this; }, i);
|
||
function verb(n) { i[n] = o[n] && function (v) { return new Promise(function (resolve, reject) { v = o[n](v), settle(resolve, reject, v.done, v.value); }); }; }
|
||
function settle(resolve, reject, d, v) { Promise.resolve(v).then(function(v) { resolve({ value: v, done: d }); }, reject); }
|
||
};
|
||
var __importStar = (this && this.__importStar) || function (mod) {
|
||
if (mod && mod.__esModule) return mod;
|
||
var result = {};
|
||
if (mod != null) for (var k in mod) if (Object.hasOwnProperty.call(mod, k)) result[k] = mod[k];
|
||
result["default"] = mod;
|
||
return result;
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const fs = __importStar(__webpack_require__(747));
|
||
const zlib = __importStar(__webpack_require__(903));
|
||
const util_1 = __webpack_require__(669);
|
||
const stat = util_1.promisify(fs.stat);
|
||
/**
|
||
* Creates a Gzip compressed file of an original file at the provided temporary filepath location
|
||
* @param {string} originalFilePath filepath of whatever will be compressed. The original file will be unmodified
|
||
* @param {string} tempFilePath the location of where the Gzip file will be created
|
||
* @returns the size of gzip file that gets created
|
||
*/
|
||
function createGZipFileOnDisk(originalFilePath, tempFilePath) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
return new Promise((resolve, reject) => {
|
||
const inputStream = fs.createReadStream(originalFilePath);
|
||
const gzip = zlib.createGzip();
|
||
const outputStream = fs.createWriteStream(tempFilePath);
|
||
inputStream.pipe(gzip).pipe(outputStream);
|
||
outputStream.on('finish', () => __awaiter(this, void 0, void 0, function* () {
|
||
// wait for stream to finish before calculating the size which is needed as part of the Content-Length header when starting an upload
|
||
const size = (yield stat(tempFilePath)).size;
|
||
resolve(size);
|
||
}));
|
||
outputStream.on('error', error => {
|
||
// eslint-disable-next-line no-console
|
||
console.log(error);
|
||
reject;
|
||
});
|
||
});
|
||
});
|
||
}
|
||
exports.createGZipFileOnDisk = createGZipFileOnDisk;
|
||
/**
|
||
* Creates a GZip file in memory using a buffer. Should be used for smaller files to reduce disk I/O
|
||
* @param originalFilePath the path to the original file that is being GZipped
|
||
* @returns a buffer with the GZip file
|
||
*/
|
||
function createGZipFileInBuffer(originalFilePath) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
return new Promise((resolve) => __awaiter(this, void 0, void 0, function* () {
|
||
var e_1, _a;
|
||
const inputStream = fs.createReadStream(originalFilePath);
|
||
const gzip = zlib.createGzip();
|
||
inputStream.pipe(gzip);
|
||
// read stream into buffer, using experimental async iterators see https://github.com/nodejs/readable-stream/issues/403#issuecomment-479069043
|
||
const chunks = [];
|
||
try {
|
||
for (var gzip_1 = __asyncValues(gzip), gzip_1_1; gzip_1_1 = yield gzip_1.next(), !gzip_1_1.done;) {
|
||
const chunk = gzip_1_1.value;
|
||
chunks.push(chunk);
|
||
}
|
||
}
|
||
catch (e_1_1) { e_1 = { error: e_1_1 }; }
|
||
finally {
|
||
try {
|
||
if (gzip_1_1 && !gzip_1_1.done && (_a = gzip_1.return)) yield _a.call(gzip_1);
|
||
}
|
||
finally { if (e_1) throw e_1.error; }
|
||
}
|
||
resolve(Buffer.concat(chunks));
|
||
}));
|
||
});
|
||
}
|
||
exports.createGZipFileInBuffer = createGZipFileInBuffer;
|
||
//# sourceMappingURL=upload-gzip.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 669:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("util");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 670:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var utils = __webpack_require__(35);
|
||
var settle = __webpack_require__(564);
|
||
var buildFullPath = __webpack_require__(960);
|
||
var buildURL = __webpack_require__(133);
|
||
var http = __webpack_require__(605);
|
||
var https = __webpack_require__(211);
|
||
var httpFollow = __webpack_require__(549).http;
|
||
var httpsFollow = __webpack_require__(549).https;
|
||
var url = __webpack_require__(835);
|
||
var zlib = __webpack_require__(903);
|
||
var pkg = __webpack_require__(361);
|
||
var createError = __webpack_require__(26);
|
||
var enhanceError = __webpack_require__(369);
|
||
|
||
var isHttps = /https:?/;
|
||
|
||
/**
|
||
*
|
||
* @param {http.ClientRequestArgs} options
|
||
* @param {AxiosProxyConfig} proxy
|
||
* @param {string} location
|
||
*/
|
||
function setProxy(options, proxy, location) {
|
||
options.hostname = proxy.host;
|
||
options.host = proxy.host;
|
||
options.port = proxy.port;
|
||
options.path = location;
|
||
|
||
// Basic proxy authorization
|
||
if (proxy.auth) {
|
||
var base64 = Buffer.from(proxy.auth.username + ':' + proxy.auth.password, 'utf8').toString('base64');
|
||
options.headers['Proxy-Authorization'] = 'Basic ' + base64;
|
||
}
|
||
|
||
// If a proxy is used, any redirects must also pass through the proxy
|
||
options.beforeRedirect = function beforeRedirect(redirection) {
|
||
redirection.headers.host = redirection.host;
|
||
setProxy(redirection, proxy, redirection.href);
|
||
};
|
||
}
|
||
|
||
/*eslint consistent-return:0*/
|
||
module.exports = function httpAdapter(config) {
|
||
return new Promise(function dispatchHttpRequest(resolvePromise, rejectPromise) {
|
||
var resolve = function resolve(value) {
|
||
resolvePromise(value);
|
||
};
|
||
var reject = function reject(value) {
|
||
rejectPromise(value);
|
||
};
|
||
var data = config.data;
|
||
var headers = config.headers;
|
||
|
||
// Set User-Agent (required by some servers)
|
||
// See https://github.com/axios/axios/issues/69
|
||
if ('User-Agent' in headers || 'user-agent' in headers) {
|
||
// User-Agent is specified; handle case where no UA header is desired
|
||
if (!headers['User-Agent'] && !headers['user-agent']) {
|
||
delete headers['User-Agent'];
|
||
delete headers['user-agent'];
|
||
}
|
||
// Otherwise, use specified value
|
||
} else {
|
||
// Only set header if it hasn't been set in config
|
||
headers['User-Agent'] = 'axios/' + pkg.version;
|
||
}
|
||
|
||
if (data && !utils.isStream(data)) {
|
||
if (Buffer.isBuffer(data)) {
|
||
// Nothing to do...
|
||
} else if (utils.isArrayBuffer(data)) {
|
||
data = Buffer.from(new Uint8Array(data));
|
||
} else if (utils.isString(data)) {
|
||
data = Buffer.from(data, 'utf-8');
|
||
} else {
|
||
return reject(createError(
|
||
'Data after transformation must be a string, an ArrayBuffer, a Buffer, or a Stream',
|
||
config
|
||
));
|
||
}
|
||
|
||
// Add Content-Length header if data exists
|
||
headers['Content-Length'] = data.length;
|
||
}
|
||
|
||
// HTTP basic authentication
|
||
var auth = undefined;
|
||
if (config.auth) {
|
||
var username = config.auth.username || '';
|
||
var password = config.auth.password || '';
|
||
auth = username + ':' + password;
|
||
}
|
||
|
||
// Parse url
|
||
var fullPath = buildFullPath(config.baseURL, config.url);
|
||
var parsed = url.parse(fullPath);
|
||
var protocol = parsed.protocol || 'http:';
|
||
|
||
if (!auth && parsed.auth) {
|
||
var urlAuth = parsed.auth.split(':');
|
||
var urlUsername = urlAuth[0] || '';
|
||
var urlPassword = urlAuth[1] || '';
|
||
auth = urlUsername + ':' + urlPassword;
|
||
}
|
||
|
||
if (auth) {
|
||
delete headers.Authorization;
|
||
}
|
||
|
||
var isHttpsRequest = isHttps.test(protocol);
|
||
var agent = isHttpsRequest ? config.httpsAgent : config.httpAgent;
|
||
|
||
var options = {
|
||
path: buildURL(parsed.path, config.params, config.paramsSerializer).replace(/^\?/, ''),
|
||
method: config.method.toUpperCase(),
|
||
headers: headers,
|
||
agent: agent,
|
||
agents: { http: config.httpAgent, https: config.httpsAgent },
|
||
auth: auth
|
||
};
|
||
|
||
if (config.socketPath) {
|
||
options.socketPath = config.socketPath;
|
||
} else {
|
||
options.hostname = parsed.hostname;
|
||
options.port = parsed.port;
|
||
}
|
||
|
||
var proxy = config.proxy;
|
||
if (!proxy && proxy !== false) {
|
||
var proxyEnv = protocol.slice(0, -1) + '_proxy';
|
||
var proxyUrl = process.env[proxyEnv] || process.env[proxyEnv.toUpperCase()];
|
||
if (proxyUrl) {
|
||
var parsedProxyUrl = url.parse(proxyUrl);
|
||
var noProxyEnv = process.env.no_proxy || process.env.NO_PROXY;
|
||
var shouldProxy = true;
|
||
|
||
if (noProxyEnv) {
|
||
var noProxy = noProxyEnv.split(',').map(function trim(s) {
|
||
return s.trim();
|
||
});
|
||
|
||
shouldProxy = !noProxy.some(function proxyMatch(proxyElement) {
|
||
if (!proxyElement) {
|
||
return false;
|
||
}
|
||
if (proxyElement === '*') {
|
||
return true;
|
||
}
|
||
if (proxyElement[0] === '.' &&
|
||
parsed.hostname.substr(parsed.hostname.length - proxyElement.length) === proxyElement) {
|
||
return true;
|
||
}
|
||
|
||
return parsed.hostname === proxyElement;
|
||
});
|
||
}
|
||
|
||
if (shouldProxy) {
|
||
proxy = {
|
||
host: parsedProxyUrl.hostname,
|
||
port: parsedProxyUrl.port,
|
||
protocol: parsedProxyUrl.protocol
|
||
};
|
||
|
||
if (parsedProxyUrl.auth) {
|
||
var proxyUrlAuth = parsedProxyUrl.auth.split(':');
|
||
proxy.auth = {
|
||
username: proxyUrlAuth[0],
|
||
password: proxyUrlAuth[1]
|
||
};
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
if (proxy) {
|
||
options.headers.host = parsed.hostname + (parsed.port ? ':' + parsed.port : '');
|
||
setProxy(options, proxy, protocol + '//' + parsed.hostname + (parsed.port ? ':' + parsed.port : '') + options.path);
|
||
}
|
||
|
||
var transport;
|
||
var isHttpsProxy = isHttpsRequest && (proxy ? isHttps.test(proxy.protocol) : true);
|
||
if (config.transport) {
|
||
transport = config.transport;
|
||
} else if (config.maxRedirects === 0) {
|
||
transport = isHttpsProxy ? https : http;
|
||
} else {
|
||
if (config.maxRedirects) {
|
||
options.maxRedirects = config.maxRedirects;
|
||
}
|
||
transport = isHttpsProxy ? httpsFollow : httpFollow;
|
||
}
|
||
|
||
if (config.maxBodyLength > -1) {
|
||
options.maxBodyLength = config.maxBodyLength;
|
||
}
|
||
|
||
// Create the request
|
||
var req = transport.request(options, function handleResponse(res) {
|
||
if (req.aborted) return;
|
||
|
||
// uncompress the response body transparently if required
|
||
var stream = res;
|
||
|
||
// return the last request in case of redirects
|
||
var lastRequest = res.req || req;
|
||
|
||
|
||
// if no content, is HEAD request or decompress disabled we should not decompress
|
||
if (res.statusCode !== 204 && lastRequest.method !== 'HEAD' && config.decompress !== false) {
|
||
switch (res.headers['content-encoding']) {
|
||
/*eslint default-case:0*/
|
||
case 'gzip':
|
||
case 'compress':
|
||
case 'deflate':
|
||
// add the unzipper to the body stream processing pipeline
|
||
stream = stream.pipe(zlib.createUnzip());
|
||
|
||
// remove the content-encoding in order to not confuse downstream operations
|
||
delete res.headers['content-encoding'];
|
||
break;
|
||
}
|
||
}
|
||
|
||
var response = {
|
||
status: res.statusCode,
|
||
statusText: res.statusMessage,
|
||
headers: res.headers,
|
||
config: config,
|
||
request: lastRequest
|
||
};
|
||
|
||
if (config.responseType === 'stream') {
|
||
response.data = stream;
|
||
settle(resolve, reject, response);
|
||
} else {
|
||
var responseBuffer = [];
|
||
var totalResponseBytes = 0;
|
||
stream.on('data', function handleStreamData(chunk) {
|
||
responseBuffer.push(chunk);
|
||
totalResponseBytes += chunk.length;
|
||
|
||
// make sure the content length is not over the maxContentLength if specified
|
||
if (config.maxContentLength > -1 && totalResponseBytes > config.maxContentLength) {
|
||
stream.destroy();
|
||
reject(createError('maxContentLength size of ' + config.maxContentLength + ' exceeded',
|
||
config, null, lastRequest));
|
||
}
|
||
});
|
||
|
||
stream.on('error', function handleStreamError(err) {
|
||
if (req.aborted) return;
|
||
reject(enhanceError(err, config, null, lastRequest));
|
||
});
|
||
|
||
stream.on('end', function handleStreamEnd() {
|
||
var responseData = Buffer.concat(responseBuffer);
|
||
if (config.responseType !== 'arraybuffer') {
|
||
responseData = responseData.toString(config.responseEncoding);
|
||
if (!config.responseEncoding || config.responseEncoding === 'utf8') {
|
||
responseData = utils.stripBOM(responseData);
|
||
}
|
||
}
|
||
|
||
response.data = responseData;
|
||
settle(resolve, reject, response);
|
||
});
|
||
}
|
||
});
|
||
|
||
// Handle errors
|
||
req.on('error', function handleRequestError(err) {
|
||
if (req.aborted && err.code !== 'ERR_FR_TOO_MANY_REDIRECTS') return;
|
||
reject(enhanceError(err, config, null, req));
|
||
});
|
||
|
||
// Handle request timeout
|
||
if (config.timeout) {
|
||
// This is forcing a int timeout to avoid problems if the `req` interface doesn't handle other types.
|
||
var timeout = parseInt(config.timeout, 10);
|
||
|
||
if (isNaN(timeout)) {
|
||
reject(createError(
|
||
'error trying to parse `config.timeout` to int',
|
||
config,
|
||
'ERR_PARSE_TIMEOUT',
|
||
req
|
||
));
|
||
|
||
return;
|
||
}
|
||
|
||
// Sometime, the response will be very slow, and does not respond, the connect event will be block by event loop system.
|
||
// And timer callback will be fired, and abort() will be invoked before connection, then get "socket hang up" and code ECONNRESET.
|
||
// At this time, if we have a large number of request, nodejs will hang up some socket on background. and the number will up and up.
|
||
// And then these socket which be hang up will devoring CPU little by little.
|
||
// ClientRequest.setTimeout will be fired on the specify milliseconds, and can make sure that abort() will be fired after connect.
|
||
req.setTimeout(timeout, function handleRequestTimeout() {
|
||
req.abort();
|
||
reject(createError(
|
||
'timeout of ' + timeout + 'ms exceeded',
|
||
config,
|
||
config.transitional && config.transitional.clarifyTimeoutError ? 'ETIMEDOUT' : 'ECONNABORTED',
|
||
req
|
||
));
|
||
});
|
||
}
|
||
|
||
if (config.cancelToken) {
|
||
// Handle cancellation
|
||
config.cancelToken.promise.then(function onCanceled(cancel) {
|
||
if (req.aborted) return;
|
||
|
||
req.abort();
|
||
reject(cancel);
|
||
});
|
||
}
|
||
|
||
// Send the request
|
||
if (utils.isStream(data)) {
|
||
data.on('error', function handleStreamError(err) {
|
||
reject(enhanceError(err, config, null, req));
|
||
}).pipe(req);
|
||
} else {
|
||
req.end(data);
|
||
}
|
||
});
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 674:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
var wrappy = __webpack_require__(11)
|
||
var reqs = Object.create(null)
|
||
var once = __webpack_require__(49)
|
||
|
||
module.exports = wrappy(inflight)
|
||
|
||
function inflight (key, cb) {
|
||
if (reqs[key]) {
|
||
reqs[key].push(cb)
|
||
return null
|
||
} else {
|
||
reqs[key] = [cb]
|
||
return makeres(key)
|
||
}
|
||
}
|
||
|
||
function makeres (key) {
|
||
return once(function RES () {
|
||
var cbs = reqs[key]
|
||
var len = cbs.length
|
||
var args = slice(arguments)
|
||
|
||
// XXX It's somewhat ambiguous whether a new callback added in this
|
||
// pass should be queued for later execution if something in the
|
||
// list of callbacks throws, or if it should just be discarded.
|
||
// However, it's such an edge case that it hardly matters, and either
|
||
// choice is likely as surprising as the other.
|
||
// As it happens, we do go ahead and schedule it for later execution.
|
||
try {
|
||
for (var i = 0; i < len; i++) {
|
||
cbs[i].apply(null, args)
|
||
}
|
||
} finally {
|
||
if (cbs.length > len) {
|
||
// added more in the interim.
|
||
// de-zalgo, just in case, but don't call again.
|
||
cbs.splice(0, len)
|
||
process.nextTick(function () {
|
||
RES.apply(null, args)
|
||
})
|
||
} else {
|
||
delete reqs[key]
|
||
}
|
||
}
|
||
})
|
||
}
|
||
|
||
function slice (args) {
|
||
var length = args.length
|
||
var array = []
|
||
|
||
for (var i = 0; i < length; i++) array[i] = args[i]
|
||
return array
|
||
}
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 681:
|
||
/***/ (function(module) {
|
||
|
||
"use strict";
|
||
|
||
|
||
function posix(path) {
|
||
return path.charAt(0) === '/';
|
||
}
|
||
|
||
function win32(path) {
|
||
// https://github.com/nodejs/node/blob/b3fcc245fb25539909ef1d5eaa01dbf92e168633/lib/path.js#L56
|
||
var splitDeviceRe = /^([a-zA-Z]:|[\\\/]{2}[^\\\/]+[\\\/]+[^\\\/]+)?([\\\/])?([\s\S]*?)$/;
|
||
var result = splitDeviceRe.exec(path);
|
||
var device = result[1] || '';
|
||
var isUnc = Boolean(device && device.charAt(1) !== ':');
|
||
|
||
// UNC paths are always absolute
|
||
return Boolean(result[2] || isUnc);
|
||
}
|
||
|
||
module.exports = process.platform === 'win32' ? win32 : posix;
|
||
module.exports.posix = posix;
|
||
module.exports.win32 = win32;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 688:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var utils = __webpack_require__(35);
|
||
|
||
module.exports = (
|
||
utils.isStandardBrowserEnv() ?
|
||
|
||
// Standard browser envs have full support of the APIs needed to test
|
||
// whether the request URL is of the same origin as current location.
|
||
(function standardBrowserEnv() {
|
||
var msie = /(msie|trident)/i.test(navigator.userAgent);
|
||
var urlParsingNode = document.createElement('a');
|
||
var originURL;
|
||
|
||
/**
|
||
* Parse a URL to discover it's components
|
||
*
|
||
* @param {String} url The URL to be parsed
|
||
* @returns {Object}
|
||
*/
|
||
function resolveURL(url) {
|
||
var href = url;
|
||
|
||
if (msie) {
|
||
// IE needs attribute set twice to normalize properties
|
||
urlParsingNode.setAttribute('href', href);
|
||
href = urlParsingNode.href;
|
||
}
|
||
|
||
urlParsingNode.setAttribute('href', href);
|
||
|
||
// urlParsingNode provides the UrlUtils interface - http://url.spec.whatwg.org/#urlutils
|
||
return {
|
||
href: urlParsingNode.href,
|
||
protocol: urlParsingNode.protocol ? urlParsingNode.protocol.replace(/:$/, '') : '',
|
||
host: urlParsingNode.host,
|
||
search: urlParsingNode.search ? urlParsingNode.search.replace(/^\?/, '') : '',
|
||
hash: urlParsingNode.hash ? urlParsingNode.hash.replace(/^#/, '') : '',
|
||
hostname: urlParsingNode.hostname,
|
||
port: urlParsingNode.port,
|
||
pathname: (urlParsingNode.pathname.charAt(0) === '/') ?
|
||
urlParsingNode.pathname :
|
||
'/' + urlParsingNode.pathname
|
||
};
|
||
}
|
||
|
||
originURL = resolveURL(window.location.href);
|
||
|
||
/**
|
||
* Determine if a URL shares the same origin as the current location
|
||
*
|
||
* @param {String} requestURL The URL to test
|
||
* @returns {boolean} True if URL shares the same origin, otherwise false
|
||
*/
|
||
return function isURLSameOrigin(requestURL) {
|
||
var parsed = (utils.isString(requestURL)) ? resolveURL(requestURL) : requestURL;
|
||
return (parsed.protocol === originURL.protocol &&
|
||
parsed.host === originURL.host);
|
||
};
|
||
})() :
|
||
|
||
// Non standard browser envs (web workers, react-native) lack needed support.
|
||
(function nonStandardBrowserEnv() {
|
||
return function isURLSameOrigin() {
|
||
return true;
|
||
};
|
||
})()
|
||
);
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 689:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
try {
|
||
var util = __webpack_require__(669);
|
||
/* istanbul ignore next */
|
||
if (typeof util.inherits !== 'function') throw '';
|
||
module.exports = util.inherits;
|
||
} catch (e) {
|
||
/* istanbul ignore next */
|
||
module.exports = __webpack_require__(315);
|
||
}
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 694:
|
||
/***/ (function(__unusedmodule, exports) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
var Inputs;
|
||
(function (Inputs) {
|
||
Inputs["Name"] = "name";
|
||
Inputs["Path"] = "path";
|
||
Inputs["Token"] = "token";
|
||
Inputs["IfNoFilesFound"] = "if-no-files-found";
|
||
Inputs["RetentionDays"] = "retention-days";
|
||
})(Inputs = exports.Inputs || (exports.Inputs = {}));
|
||
var NoFileOptions;
|
||
(function (NoFileOptions) {
|
||
/**
|
||
* Default. Output a warning but do not fail the action
|
||
*/
|
||
NoFileOptions["warn"] = "warn";
|
||
/**
|
||
* Fail the action with an error message
|
||
*/
|
||
NoFileOptions["error"] = "error";
|
||
/**
|
||
* Do not output any warnings or errors, the action does not fail
|
||
*/
|
||
NoFileOptions["ignore"] = "ignore";
|
||
})(NoFileOptions = exports.NoFileOptions || (exports.NoFileOptions = {}));
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 727:
|
||
/***/ (function(module) {
|
||
|
||
"use strict";
|
||
|
||
|
||
module.exports = function bind(fn, thisArg) {
|
||
return function wrap() {
|
||
var args = new Array(arguments.length);
|
||
for (var i = 0; i < args.length; i++) {
|
||
args[i] = arguments[i];
|
||
}
|
||
return fn.apply(thisArg, args);
|
||
};
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 728:
|
||
/***/ (function(__unusedmodule, exports) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
class SearchState {
|
||
constructor(path, level) {
|
||
this.path = path;
|
||
this.level = level;
|
||
}
|
||
}
|
||
exports.SearchState = SearchState;
|
||
//# sourceMappingURL=internal-search-state.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 732:
|
||
/***/ (function(module) {
|
||
|
||
"use strict";
|
||
|
||
|
||
module.exports = function isCancel(value) {
|
||
return !!(value && value.__CANCEL__);
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 747:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("fs");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 761:
|
||
/***/ (function(module) {
|
||
|
||
/**
|
||
* Helpers.
|
||
*/
|
||
|
||
var s = 1000;
|
||
var m = s * 60;
|
||
var h = m * 60;
|
||
var d = h * 24;
|
||
var w = d * 7;
|
||
var y = d * 365.25;
|
||
|
||
/**
|
||
* Parse or format the given `val`.
|
||
*
|
||
* Options:
|
||
*
|
||
* - `long` verbose formatting [false]
|
||
*
|
||
* @param {String|Number} val
|
||
* @param {Object} [options]
|
||
* @throws {Error} throw an error if val is not a non-empty string or a number
|
||
* @return {String|Number}
|
||
* @api public
|
||
*/
|
||
|
||
module.exports = function(val, options) {
|
||
options = options || {};
|
||
var type = typeof val;
|
||
if (type === 'string' && val.length > 0) {
|
||
return parse(val);
|
||
} else if (type === 'number' && isFinite(val)) {
|
||
return options.long ? fmtLong(val) : fmtShort(val);
|
||
}
|
||
throw new Error(
|
||
'val is not a non-empty string or a valid number. val=' +
|
||
JSON.stringify(val)
|
||
);
|
||
};
|
||
|
||
/**
|
||
* Parse the given `str` and return milliseconds.
|
||
*
|
||
* @param {String} str
|
||
* @return {Number}
|
||
* @api private
|
||
*/
|
||
|
||
function parse(str) {
|
||
str = String(str);
|
||
if (str.length > 100) {
|
||
return;
|
||
}
|
||
var match = /^(-?(?:\d+)?\.?\d+) *(milliseconds?|msecs?|ms|seconds?|secs?|s|minutes?|mins?|m|hours?|hrs?|h|days?|d|weeks?|w|years?|yrs?|y)?$/i.exec(
|
||
str
|
||
);
|
||
if (!match) {
|
||
return;
|
||
}
|
||
var n = parseFloat(match[1]);
|
||
var type = (match[2] || 'ms').toLowerCase();
|
||
switch (type) {
|
||
case 'years':
|
||
case 'year':
|
||
case 'yrs':
|
||
case 'yr':
|
||
case 'y':
|
||
return n * y;
|
||
case 'weeks':
|
||
case 'week':
|
||
case 'w':
|
||
return n * w;
|
||
case 'days':
|
||
case 'day':
|
||
case 'd':
|
||
return n * d;
|
||
case 'hours':
|
||
case 'hour':
|
||
case 'hrs':
|
||
case 'hr':
|
||
case 'h':
|
||
return n * h;
|
||
case 'minutes':
|
||
case 'minute':
|
||
case 'mins':
|
||
case 'min':
|
||
case 'm':
|
||
return n * m;
|
||
case 'seconds':
|
||
case 'second':
|
||
case 'secs':
|
||
case 'sec':
|
||
case 's':
|
||
return n * s;
|
||
case 'milliseconds':
|
||
case 'millisecond':
|
||
case 'msecs':
|
||
case 'msec':
|
||
case 'ms':
|
||
return n;
|
||
default:
|
||
return undefined;
|
||
}
|
||
}
|
||
|
||
/**
|
||
* Short format for `ms`.
|
||
*
|
||
* @param {Number} ms
|
||
* @return {String}
|
||
* @api private
|
||
*/
|
||
|
||
function fmtShort(ms) {
|
||
var msAbs = Math.abs(ms);
|
||
if (msAbs >= d) {
|
||
return Math.round(ms / d) + 'd';
|
||
}
|
||
if (msAbs >= h) {
|
||
return Math.round(ms / h) + 'h';
|
||
}
|
||
if (msAbs >= m) {
|
||
return Math.round(ms / m) + 'm';
|
||
}
|
||
if (msAbs >= s) {
|
||
return Math.round(ms / s) + 's';
|
||
}
|
||
return ms + 'ms';
|
||
}
|
||
|
||
/**
|
||
* Long format for `ms`.
|
||
*
|
||
* @param {Number} ms
|
||
* @return {String}
|
||
* @api private
|
||
*/
|
||
|
||
function fmtLong(ms) {
|
||
var msAbs = Math.abs(ms);
|
||
if (msAbs >= d) {
|
||
return plural(ms, msAbs, d, 'day');
|
||
}
|
||
if (msAbs >= h) {
|
||
return plural(ms, msAbs, h, 'hour');
|
||
}
|
||
if (msAbs >= m) {
|
||
return plural(ms, msAbs, m, 'minute');
|
||
}
|
||
if (msAbs >= s) {
|
||
return plural(ms, msAbs, s, 'second');
|
||
}
|
||
return ms + ' ms';
|
||
}
|
||
|
||
/**
|
||
* Pluralization helper.
|
||
*/
|
||
|
||
function plural(ms, msAbs, n, name) {
|
||
var isPlural = msAbs >= n * 1.5;
|
||
return Math.round(ms / n) + ' ' + name + (isPlural ? 's' : '');
|
||
}
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 779:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var utils = __webpack_require__(35);
|
||
var buildURL = __webpack_require__(133);
|
||
var InterceptorManager = __webpack_require__(283);
|
||
var dispatchRequest = __webpack_require__(946);
|
||
var mergeConfig = __webpack_require__(825);
|
||
var validator = __webpack_require__(106);
|
||
|
||
var validators = validator.validators;
|
||
/**
|
||
* Create a new instance of Axios
|
||
*
|
||
* @param {Object} instanceConfig The default config for the instance
|
||
*/
|
||
function Axios(instanceConfig) {
|
||
this.defaults = instanceConfig;
|
||
this.interceptors = {
|
||
request: new InterceptorManager(),
|
||
response: new InterceptorManager()
|
||
};
|
||
}
|
||
|
||
/**
|
||
* Dispatch a request
|
||
*
|
||
* @param {Object} config The config specific for this request (merged with this.defaults)
|
||
*/
|
||
Axios.prototype.request = function request(config) {
|
||
/*eslint no-param-reassign:0*/
|
||
// Allow for axios('example/url'[, config]) a la fetch API
|
||
if (typeof config === 'string') {
|
||
config = arguments[1] || {};
|
||
config.url = arguments[0];
|
||
} else {
|
||
config = config || {};
|
||
}
|
||
|
||
config = mergeConfig(this.defaults, config);
|
||
|
||
// Set config.method
|
||
if (config.method) {
|
||
config.method = config.method.toLowerCase();
|
||
} else if (this.defaults.method) {
|
||
config.method = this.defaults.method.toLowerCase();
|
||
} else {
|
||
config.method = 'get';
|
||
}
|
||
|
||
var transitional = config.transitional;
|
||
|
||
if (transitional !== undefined) {
|
||
validator.assertOptions(transitional, {
|
||
silentJSONParsing: validators.transitional(validators.boolean, '1.0.0'),
|
||
forcedJSONParsing: validators.transitional(validators.boolean, '1.0.0'),
|
||
clarifyTimeoutError: validators.transitional(validators.boolean, '1.0.0')
|
||
}, false);
|
||
}
|
||
|
||
// filter out skipped interceptors
|
||
var requestInterceptorChain = [];
|
||
var synchronousRequestInterceptors = true;
|
||
this.interceptors.request.forEach(function unshiftRequestInterceptors(interceptor) {
|
||
if (typeof interceptor.runWhen === 'function' && interceptor.runWhen(config) === false) {
|
||
return;
|
||
}
|
||
|
||
synchronousRequestInterceptors = synchronousRequestInterceptors && interceptor.synchronous;
|
||
|
||
requestInterceptorChain.unshift(interceptor.fulfilled, interceptor.rejected);
|
||
});
|
||
|
||
var responseInterceptorChain = [];
|
||
this.interceptors.response.forEach(function pushResponseInterceptors(interceptor) {
|
||
responseInterceptorChain.push(interceptor.fulfilled, interceptor.rejected);
|
||
});
|
||
|
||
var promise;
|
||
|
||
if (!synchronousRequestInterceptors) {
|
||
var chain = [dispatchRequest, undefined];
|
||
|
||
Array.prototype.unshift.apply(chain, requestInterceptorChain);
|
||
chain = chain.concat(responseInterceptorChain);
|
||
|
||
promise = Promise.resolve(config);
|
||
while (chain.length) {
|
||
promise = promise.then(chain.shift(), chain.shift());
|
||
}
|
||
|
||
return promise;
|
||
}
|
||
|
||
|
||
var newConfig = config;
|
||
while (requestInterceptorChain.length) {
|
||
var onFulfilled = requestInterceptorChain.shift();
|
||
var onRejected = requestInterceptorChain.shift();
|
||
try {
|
||
newConfig = onFulfilled(newConfig);
|
||
} catch (error) {
|
||
onRejected(error);
|
||
break;
|
||
}
|
||
}
|
||
|
||
try {
|
||
promise = dispatchRequest(newConfig);
|
||
} catch (error) {
|
||
return Promise.reject(error);
|
||
}
|
||
|
||
while (responseInterceptorChain.length) {
|
||
promise = promise.then(responseInterceptorChain.shift(), responseInterceptorChain.shift());
|
||
}
|
||
|
||
return promise;
|
||
};
|
||
|
||
Axios.prototype.getUri = function getUri(config) {
|
||
config = mergeConfig(this.defaults, config);
|
||
return buildURL(config.url, config.params, config.paramsSerializer).replace(/^\?/, '');
|
||
};
|
||
|
||
// Provide aliases for supported request methods
|
||
utils.forEach(['delete', 'get', 'head', 'options'], function forEachMethodNoData(method) {
|
||
/*eslint func-names:0*/
|
||
Axios.prototype[method] = function(url, config) {
|
||
return this.request(mergeConfig(config || {}, {
|
||
method: method,
|
||
url: url,
|
||
data: (config || {}).data
|
||
}));
|
||
};
|
||
});
|
||
|
||
utils.forEach(['post', 'put', 'patch'], function forEachMethodWithData(method) {
|
||
/*eslint func-names:0*/
|
||
Axios.prototype[method] = function(url, data, config) {
|
||
return this.request(mergeConfig(config || {}, {
|
||
method: method,
|
||
url: url,
|
||
data: data
|
||
}));
|
||
};
|
||
});
|
||
|
||
module.exports = Axios;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 784:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
/**
|
||
* Detect Electron renderer / nwjs process, which is node, but we should
|
||
* treat as a browser.
|
||
*/
|
||
|
||
if (typeof process === 'undefined' || process.type === 'renderer' || process.browser === true || process.__nwjs) {
|
||
module.exports = __webpack_require__(794);
|
||
} else {
|
||
module.exports = __webpack_require__(81);
|
||
}
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 794:
|
||
/***/ (function(module, exports, __webpack_require__) {
|
||
|
||
/* eslint-env browser */
|
||
|
||
/**
|
||
* This is the web browser implementation of `debug()`.
|
||
*/
|
||
|
||
exports.log = log;
|
||
exports.formatArgs = formatArgs;
|
||
exports.save = save;
|
||
exports.load = load;
|
||
exports.useColors = useColors;
|
||
exports.storage = localstorage();
|
||
|
||
/**
|
||
* Colors.
|
||
*/
|
||
|
||
exports.colors = [
|
||
'#0000CC',
|
||
'#0000FF',
|
||
'#0033CC',
|
||
'#0033FF',
|
||
'#0066CC',
|
||
'#0066FF',
|
||
'#0099CC',
|
||
'#0099FF',
|
||
'#00CC00',
|
||
'#00CC33',
|
||
'#00CC66',
|
||
'#00CC99',
|
||
'#00CCCC',
|
||
'#00CCFF',
|
||
'#3300CC',
|
||
'#3300FF',
|
||
'#3333CC',
|
||
'#3333FF',
|
||
'#3366CC',
|
||
'#3366FF',
|
||
'#3399CC',
|
||
'#3399FF',
|
||
'#33CC00',
|
||
'#33CC33',
|
||
'#33CC66',
|
||
'#33CC99',
|
||
'#33CCCC',
|
||
'#33CCFF',
|
||
'#6600CC',
|
||
'#6600FF',
|
||
'#6633CC',
|
||
'#6633FF',
|
||
'#66CC00',
|
||
'#66CC33',
|
||
'#9900CC',
|
||
'#9900FF',
|
||
'#9933CC',
|
||
'#9933FF',
|
||
'#99CC00',
|
||
'#99CC33',
|
||
'#CC0000',
|
||
'#CC0033',
|
||
'#CC0066',
|
||
'#CC0099',
|
||
'#CC00CC',
|
||
'#CC00FF',
|
||
'#CC3300',
|
||
'#CC3333',
|
||
'#CC3366',
|
||
'#CC3399',
|
||
'#CC33CC',
|
||
'#CC33FF',
|
||
'#CC6600',
|
||
'#CC6633',
|
||
'#CC9900',
|
||
'#CC9933',
|
||
'#CCCC00',
|
||
'#CCCC33',
|
||
'#FF0000',
|
||
'#FF0033',
|
||
'#FF0066',
|
||
'#FF0099',
|
||
'#FF00CC',
|
||
'#FF00FF',
|
||
'#FF3300',
|
||
'#FF3333',
|
||
'#FF3366',
|
||
'#FF3399',
|
||
'#FF33CC',
|
||
'#FF33FF',
|
||
'#FF6600',
|
||
'#FF6633',
|
||
'#FF9900',
|
||
'#FF9933',
|
||
'#FFCC00',
|
||
'#FFCC33'
|
||
];
|
||
|
||
/**
|
||
* Currently only WebKit-based Web Inspectors, Firefox >= v31,
|
||
* and the Firebug extension (any Firefox version) are known
|
||
* to support "%c" CSS customizations.
|
||
*
|
||
* TODO: add a `localStorage` variable to explicitly enable/disable colors
|
||
*/
|
||
|
||
// eslint-disable-next-line complexity
|
||
function useColors() {
|
||
// NB: In an Electron preload script, document will be defined but not fully
|
||
// initialized. Since we know we're in Chrome, we'll just detect this case
|
||
// explicitly
|
||
if (typeof window !== 'undefined' && window.process && (window.process.type === 'renderer' || window.process.__nwjs)) {
|
||
return true;
|
||
}
|
||
|
||
// Internet Explorer and Edge do not support colors.
|
||
if (typeof navigator !== 'undefined' && navigator.userAgent && navigator.userAgent.toLowerCase().match(/(edge|trident)\/(\d+)/)) {
|
||
return false;
|
||
}
|
||
|
||
// Is webkit? http://stackoverflow.com/a/16459606/376773
|
||
// document is undefined in react-native: https://github.com/facebook/react-native/pull/1632
|
||
return (typeof document !== 'undefined' && document.documentElement && document.documentElement.style && document.documentElement.style.WebkitAppearance) ||
|
||
// Is firebug? http://stackoverflow.com/a/398120/376773
|
||
(typeof window !== 'undefined' && window.console && (window.console.firebug || (window.console.exception && window.console.table))) ||
|
||
// Is firefox >= v31?
|
||
// https://developer.mozilla.org/en-US/docs/Tools/Web_Console#Styling_messages
|
||
(typeof navigator !== 'undefined' && navigator.userAgent && navigator.userAgent.toLowerCase().match(/firefox\/(\d+)/) && parseInt(RegExp.$1, 10) >= 31) ||
|
||
// Double check webkit in userAgent just in case we are in a worker
|
||
(typeof navigator !== 'undefined' && navigator.userAgent && navigator.userAgent.toLowerCase().match(/applewebkit\/(\d+)/));
|
||
}
|
||
|
||
/**
|
||
* Colorize log arguments if enabled.
|
||
*
|
||
* @api public
|
||
*/
|
||
|
||
function formatArgs(args) {
|
||
args[0] = (this.useColors ? '%c' : '') +
|
||
this.namespace +
|
||
(this.useColors ? ' %c' : ' ') +
|
||
args[0] +
|
||
(this.useColors ? '%c ' : ' ') +
|
||
'+' + module.exports.humanize(this.diff);
|
||
|
||
if (!this.useColors) {
|
||
return;
|
||
}
|
||
|
||
const c = 'color: ' + this.color;
|
||
args.splice(1, 0, c, 'color: inherit');
|
||
|
||
// The final "%c" is somewhat tricky, because there could be other
|
||
// arguments passed either before or after the %c, so we need to
|
||
// figure out the correct index to insert the CSS into
|
||
let index = 0;
|
||
let lastC = 0;
|
||
args[0].replace(/%[a-zA-Z%]/g, match => {
|
||
if (match === '%%') {
|
||
return;
|
||
}
|
||
index++;
|
||
if (match === '%c') {
|
||
// We only are interested in the *last* %c
|
||
// (the user may have provided their own)
|
||
lastC = index;
|
||
}
|
||
});
|
||
|
||
args.splice(lastC, 0, c);
|
||
}
|
||
|
||
/**
|
||
* Invokes `console.log()` when available.
|
||
* No-op when `console.log` is not a "function".
|
||
*
|
||
* @api public
|
||
*/
|
||
function log(...args) {
|
||
// This hackery is required for IE8/9, where
|
||
// the `console.log` function doesn't have 'apply'
|
||
return typeof console === 'object' &&
|
||
console.log &&
|
||
console.log(...args);
|
||
}
|
||
|
||
/**
|
||
* Save `namespaces`.
|
||
*
|
||
* @param {String} namespaces
|
||
* @api private
|
||
*/
|
||
function save(namespaces) {
|
||
try {
|
||
if (namespaces) {
|
||
exports.storage.setItem('debug', namespaces);
|
||
} else {
|
||
exports.storage.removeItem('debug');
|
||
}
|
||
} catch (error) {
|
||
// Swallow
|
||
// XXX (@Qix-) should we be logging these?
|
||
}
|
||
}
|
||
|
||
/**
|
||
* Load `namespaces`.
|
||
*
|
||
* @return {String} returns the previously persisted debug modes
|
||
* @api private
|
||
*/
|
||
function load() {
|
||
let r;
|
||
try {
|
||
r = exports.storage.getItem('debug');
|
||
} catch (error) {
|
||
// Swallow
|
||
// XXX (@Qix-) should we be logging these?
|
||
}
|
||
|
||
// If debug isn't set in LS, and we're in Electron, try to load $DEBUG
|
||
if (!r && typeof process !== 'undefined' && 'env' in process) {
|
||
r = process.env.DEBUG;
|
||
}
|
||
|
||
return r;
|
||
}
|
||
|
||
/**
|
||
* Localstorage attempts to return the localstorage.
|
||
*
|
||
* This is necessary because safari throws
|
||
* when a user disables cookies/localstorage
|
||
* and you attempt to access it.
|
||
*
|
||
* @return {LocalStorage}
|
||
* @api private
|
||
*/
|
||
|
||
function localstorage() {
|
||
try {
|
||
// TVMLKit (Apple TV JS Runtime) does not have a window object, just localStorage in the global context
|
||
// The Browser also has localStorage in the global context.
|
||
return localStorage;
|
||
} catch (error) {
|
||
// Swallow
|
||
// XXX (@Qix-) should we be logging these?
|
||
}
|
||
}
|
||
|
||
module.exports = __webpack_require__(486)(exports);
|
||
|
||
const {formatters} = module.exports;
|
||
|
||
/**
|
||
* Map %j to `JSON.stringify()`, since no Web Inspectors do that by default.
|
||
*/
|
||
|
||
formatters.j = function (v) {
|
||
try {
|
||
return JSON.stringify(v);
|
||
} catch (error) {
|
||
return '[UnexpectedJSONParseError]: ' + error.message;
|
||
}
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 825:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var utils = __webpack_require__(35);
|
||
|
||
/**
|
||
* Config-specific merge-function which creates a new config-object
|
||
* by merging two configuration objects together.
|
||
*
|
||
* @param {Object} config1
|
||
* @param {Object} config2
|
||
* @returns {Object} New object resulting from merging config2 to config1
|
||
*/
|
||
module.exports = function mergeConfig(config1, config2) {
|
||
// eslint-disable-next-line no-param-reassign
|
||
config2 = config2 || {};
|
||
var config = {};
|
||
|
||
var valueFromConfig2Keys = ['url', 'method', 'data'];
|
||
var mergeDeepPropertiesKeys = ['headers', 'auth', 'proxy', 'params'];
|
||
var defaultToConfig2Keys = [
|
||
'baseURL', 'transformRequest', 'transformResponse', 'paramsSerializer',
|
||
'timeout', 'timeoutMessage', 'withCredentials', 'adapter', 'responseType', 'xsrfCookieName',
|
||
'xsrfHeaderName', 'onUploadProgress', 'onDownloadProgress', 'decompress',
|
||
'maxContentLength', 'maxBodyLength', 'maxRedirects', 'transport', 'httpAgent',
|
||
'httpsAgent', 'cancelToken', 'socketPath', 'responseEncoding'
|
||
];
|
||
var directMergeKeys = ['validateStatus'];
|
||
|
||
function getMergedValue(target, source) {
|
||
if (utils.isPlainObject(target) && utils.isPlainObject(source)) {
|
||
return utils.merge(target, source);
|
||
} else if (utils.isPlainObject(source)) {
|
||
return utils.merge({}, source);
|
||
} else if (utils.isArray(source)) {
|
||
return source.slice();
|
||
}
|
||
return source;
|
||
}
|
||
|
||
function mergeDeepProperties(prop) {
|
||
if (!utils.isUndefined(config2[prop])) {
|
||
config[prop] = getMergedValue(config1[prop], config2[prop]);
|
||
} else if (!utils.isUndefined(config1[prop])) {
|
||
config[prop] = getMergedValue(undefined, config1[prop]);
|
||
}
|
||
}
|
||
|
||
utils.forEach(valueFromConfig2Keys, function valueFromConfig2(prop) {
|
||
if (!utils.isUndefined(config2[prop])) {
|
||
config[prop] = getMergedValue(undefined, config2[prop]);
|
||
}
|
||
});
|
||
|
||
utils.forEach(mergeDeepPropertiesKeys, mergeDeepProperties);
|
||
|
||
utils.forEach(defaultToConfig2Keys, function defaultToConfig2(prop) {
|
||
if (!utils.isUndefined(config2[prop])) {
|
||
config[prop] = getMergedValue(undefined, config2[prop]);
|
||
} else if (!utils.isUndefined(config1[prop])) {
|
||
config[prop] = getMergedValue(undefined, config1[prop]);
|
||
}
|
||
});
|
||
|
||
utils.forEach(directMergeKeys, function merge(prop) {
|
||
if (prop in config2) {
|
||
config[prop] = getMergedValue(config1[prop], config2[prop]);
|
||
} else if (prop in config1) {
|
||
config[prop] = getMergedValue(undefined, config1[prop]);
|
||
}
|
||
});
|
||
|
||
var axiosKeys = valueFromConfig2Keys
|
||
.concat(mergeDeepPropertiesKeys)
|
||
.concat(defaultToConfig2Keys)
|
||
.concat(directMergeKeys);
|
||
|
||
var otherKeys = Object
|
||
.keys(config1)
|
||
.concat(Object.keys(config2))
|
||
.filter(function filterAxiosKeys(key) {
|
||
return axiosKeys.indexOf(key) === -1;
|
||
});
|
||
|
||
utils.forEach(otherKeys, mergeDeepProperties);
|
||
|
||
return config;
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 826:
|
||
/***/ (function(module) {
|
||
|
||
"use strict";
|
||
|
||
|
||
/**
|
||
* A `Cancel` is an object that is thrown when an operation is canceled.
|
||
*
|
||
* @class
|
||
* @param {string=} message The message.
|
||
*/
|
||
function Cancel(message) {
|
||
this.message = message;
|
||
}
|
||
|
||
Cancel.prototype.toString = function toString() {
|
||
return 'Cancel' + (this.message ? ': ' + this.message : '');
|
||
};
|
||
|
||
Cancel.prototype.__CANCEL__ = true;
|
||
|
||
module.exports = Cancel;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 835:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("url");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 855:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
||
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
||
return new (P || (P = Promise))(function (resolve, reject) {
|
||
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
||
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
||
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
||
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
||
});
|
||
};
|
||
var __importStar = (this && this.__importStar) || function (mod) {
|
||
if (mod && mod.__esModule) return mod;
|
||
var result = {};
|
||
if (mod != null) for (var k in mod) if (Object.hasOwnProperty.call(mod, k)) result[k] = mod[k];
|
||
result["default"] = mod;
|
||
return result;
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const fs = __importStar(__webpack_require__(747));
|
||
const core = __importStar(__webpack_require__(470));
|
||
const zlib = __importStar(__webpack_require__(903));
|
||
const utils_1 = __webpack_require__(870);
|
||
const url_1 = __webpack_require__(835);
|
||
const status_reporter_1 = __webpack_require__(176);
|
||
const perf_hooks_1 = __webpack_require__(630);
|
||
const http_manager_1 = __webpack_require__(452);
|
||
const config_variables_1 = __webpack_require__(401);
|
||
const requestUtils_1 = __webpack_require__(489);
|
||
class DownloadHttpClient {
|
||
constructor() {
|
||
this.downloadHttpManager = new http_manager_1.HttpManager(config_variables_1.getDownloadFileConcurrency(), '@actions/artifact-download');
|
||
// downloads are usually significantly faster than uploads so display status information every second
|
||
this.statusReporter = new status_reporter_1.StatusReporter(1000);
|
||
}
|
||
/**
|
||
* Gets a list of all artifacts that are in a specific container
|
||
*/
|
||
listArtifacts() {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
const artifactUrl = utils_1.getArtifactUrl();
|
||
// use the first client from the httpManager, `keep-alive` is not used so the connection will close immediately
|
||
const client = this.downloadHttpManager.getClient(0);
|
||
const headers = utils_1.getDownloadHeaders('application/json');
|
||
const response = yield requestUtils_1.retryHttpClientRequest('List Artifacts', () => __awaiter(this, void 0, void 0, function* () { return client.get(artifactUrl, headers); }));
|
||
const body = yield response.readBody();
|
||
return JSON.parse(body);
|
||
});
|
||
}
|
||
/**
|
||
* Fetches a set of container items that describe the contents of an artifact
|
||
* @param artifactName the name of the artifact
|
||
* @param containerUrl the artifact container URL for the run
|
||
*/
|
||
getContainerItems(artifactName, containerUrl) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
// the itemPath search parameter controls which containers will be returned
|
||
const resourceUrl = new url_1.URL(containerUrl);
|
||
resourceUrl.searchParams.append('itemPath', artifactName);
|
||
// use the first client from the httpManager, `keep-alive` is not used so the connection will close immediately
|
||
const client = this.downloadHttpManager.getClient(0);
|
||
const headers = utils_1.getDownloadHeaders('application/json');
|
||
const response = yield requestUtils_1.retryHttpClientRequest('Get Container Items', () => __awaiter(this, void 0, void 0, function* () { return client.get(resourceUrl.toString(), headers); }));
|
||
const body = yield response.readBody();
|
||
return JSON.parse(body);
|
||
});
|
||
}
|
||
/**
|
||
* Concurrently downloads all the files that are part of an artifact
|
||
* @param downloadItems information about what items to download and where to save them
|
||
*/
|
||
downloadSingleArtifact(downloadItems) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
const DOWNLOAD_CONCURRENCY = config_variables_1.getDownloadFileConcurrency();
|
||
// limit the number of files downloaded at a single time
|
||
core.debug(`Download file concurrency is set to ${DOWNLOAD_CONCURRENCY}`);
|
||
const parallelDownloads = [...new Array(DOWNLOAD_CONCURRENCY).keys()];
|
||
let currentFile = 0;
|
||
let downloadedFiles = 0;
|
||
core.info(`Total number of files that will be downloaded: ${downloadItems.length}`);
|
||
this.statusReporter.setTotalNumberOfFilesToProcess(downloadItems.length);
|
||
this.statusReporter.start();
|
||
yield Promise.all(parallelDownloads.map((index) => __awaiter(this, void 0, void 0, function* () {
|
||
while (currentFile < downloadItems.length) {
|
||
const currentFileToDownload = downloadItems[currentFile];
|
||
currentFile += 1;
|
||
const startTime = perf_hooks_1.performance.now();
|
||
yield this.downloadIndividualFile(index, currentFileToDownload.sourceLocation, currentFileToDownload.targetPath);
|
||
if (core.isDebug()) {
|
||
core.debug(`File: ${++downloadedFiles}/${downloadItems.length}. ${currentFileToDownload.targetPath} took ${(perf_hooks_1.performance.now() - startTime).toFixed(3)} milliseconds to finish downloading`);
|
||
}
|
||
this.statusReporter.incrementProcessedCount();
|
||
}
|
||
})))
|
||
.catch(error => {
|
||
throw new Error(`Unable to download the artifact: ${error}`);
|
||
})
|
||
.finally(() => {
|
||
this.statusReporter.stop();
|
||
// safety dispose all connections
|
||
this.downloadHttpManager.disposeAndReplaceAllClients();
|
||
});
|
||
});
|
||
}
|
||
/**
|
||
* Downloads an individual file
|
||
* @param httpClientIndex the index of the http client that is used to make all of the calls
|
||
* @param artifactLocation origin location where a file will be downloaded from
|
||
* @param downloadPath destination location for the file being downloaded
|
||
*/
|
||
downloadIndividualFile(httpClientIndex, artifactLocation, downloadPath) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
let retryCount = 0;
|
||
const retryLimit = config_variables_1.getRetryLimit();
|
||
let destinationStream = fs.createWriteStream(downloadPath);
|
||
const headers = utils_1.getDownloadHeaders('application/json', true, true);
|
||
// a single GET request is used to download a file
|
||
const makeDownloadRequest = () => __awaiter(this, void 0, void 0, function* () {
|
||
const client = this.downloadHttpManager.getClient(httpClientIndex);
|
||
return yield client.get(artifactLocation, headers);
|
||
});
|
||
// check the response headers to determine if the file was compressed using gzip
|
||
const isGzip = (incomingHeaders) => {
|
||
return ('content-encoding' in incomingHeaders &&
|
||
incomingHeaders['content-encoding'] === 'gzip');
|
||
};
|
||
// Increments the current retry count and then checks if the retry limit has been reached
|
||
// If there have been too many retries, fail so the download stops. If there is a retryAfterValue value provided,
|
||
// it will be used
|
||
const backOff = (retryAfterValue) => __awaiter(this, void 0, void 0, function* () {
|
||
retryCount++;
|
||
if (retryCount > retryLimit) {
|
||
return Promise.reject(new Error(`Retry limit has been reached. Unable to download ${artifactLocation}`));
|
||
}
|
||
else {
|
||
this.downloadHttpManager.disposeAndReplaceClient(httpClientIndex);
|
||
if (retryAfterValue) {
|
||
// Back off by waiting the specified time denoted by the retry-after header
|
||
core.info(`Backoff due to too many requests, retry #${retryCount}. Waiting for ${retryAfterValue} milliseconds before continuing the download`);
|
||
yield utils_1.sleep(retryAfterValue);
|
||
}
|
||
else {
|
||
// Back off using an exponential value that depends on the retry count
|
||
const backoffTime = utils_1.getExponentialRetryTimeInMilliseconds(retryCount);
|
||
core.info(`Exponential backoff for retry #${retryCount}. Waiting for ${backoffTime} milliseconds before continuing the download`);
|
||
yield utils_1.sleep(backoffTime);
|
||
}
|
||
core.info(`Finished backoff for retry #${retryCount}, continuing with download`);
|
||
}
|
||
});
|
||
const isAllBytesReceived = (expected, received) => {
|
||
// be lenient, if any input is missing, assume success, i.e. not truncated
|
||
if (!expected ||
|
||
!received ||
|
||
process.env['ACTIONS_ARTIFACT_SKIP_DOWNLOAD_VALIDATION']) {
|
||
core.info('Skipping download validation.');
|
||
return true;
|
||
}
|
||
return parseInt(expected) === received;
|
||
};
|
||
const resetDestinationStream = (fileDownloadPath) => __awaiter(this, void 0, void 0, function* () {
|
||
destinationStream.close();
|
||
yield utils_1.rmFile(fileDownloadPath);
|
||
destinationStream = fs.createWriteStream(fileDownloadPath);
|
||
});
|
||
// keep trying to download a file until a retry limit has been reached
|
||
while (retryCount <= retryLimit) {
|
||
let response;
|
||
try {
|
||
response = yield makeDownloadRequest();
|
||
if (core.isDebug()) {
|
||
utils_1.displayHttpDiagnostics(response);
|
||
}
|
||
}
|
||
catch (error) {
|
||
// if an error is caught, it is usually indicative of a timeout so retry the download
|
||
core.info('An error occurred while attempting to download a file');
|
||
// eslint-disable-next-line no-console
|
||
console.log(error);
|
||
// increment the retryCount and use exponential backoff to wait before making the next request
|
||
yield backOff();
|
||
continue;
|
||
}
|
||
let forceRetry = false;
|
||
if (utils_1.isSuccessStatusCode(response.message.statusCode)) {
|
||
// The body contains the contents of the file however calling response.readBody() causes all the content to be converted to a string
|
||
// which can cause some gzip encoded data to be lost
|
||
// Instead of using response.readBody(), response.message is a readableStream that can be directly used to get the raw body contents
|
||
try {
|
||
const isGzipped = isGzip(response.message.headers);
|
||
yield this.pipeResponseToFile(response, destinationStream, isGzipped);
|
||
if (isGzipped ||
|
||
isAllBytesReceived(response.message.headers['content-length'], yield utils_1.getFileSize(downloadPath))) {
|
||
return;
|
||
}
|
||
else {
|
||
forceRetry = true;
|
||
}
|
||
}
|
||
catch (error) {
|
||
// retry on error, most likely streams were corrupted
|
||
forceRetry = true;
|
||
}
|
||
}
|
||
if (forceRetry || utils_1.isRetryableStatusCode(response.message.statusCode)) {
|
||
core.info(`A ${response.message.statusCode} response code has been received while attempting to download an artifact`);
|
||
resetDestinationStream(downloadPath);
|
||
// if a throttled status code is received, try to get the retryAfter header value, else differ to standard exponential backoff
|
||
utils_1.isThrottledStatusCode(response.message.statusCode)
|
||
? yield backOff(utils_1.tryGetRetryAfterValueTimeInMilliseconds(response.message.headers))
|
||
: yield backOff();
|
||
}
|
||
else {
|
||
// Some unexpected response code, fail immediately and stop the download
|
||
utils_1.displayHttpDiagnostics(response);
|
||
return Promise.reject(new Error(`Unexpected http ${response.message.statusCode} during download for ${artifactLocation}`));
|
||
}
|
||
}
|
||
});
|
||
}
|
||
/**
|
||
* Pipes the response from downloading an individual file to the appropriate destination stream while decoding gzip content if necessary
|
||
* @param response the http response received when downloading a file
|
||
* @param destinationStream the stream where the file should be written to
|
||
* @param isGzip a boolean denoting if the content is compressed using gzip and if we need to decode it
|
||
*/
|
||
pipeResponseToFile(response, destinationStream, isGzip) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
yield new Promise((resolve, reject) => {
|
||
if (isGzip) {
|
||
const gunzip = zlib.createGunzip();
|
||
response.message
|
||
.on('error', error => {
|
||
core.error(`An error occurred while attempting to read the response stream`);
|
||
gunzip.close();
|
||
destinationStream.close();
|
||
reject(error);
|
||
})
|
||
.pipe(gunzip)
|
||
.on('error', error => {
|
||
core.error(`An error occurred while attempting to decompress the response stream`);
|
||
destinationStream.close();
|
||
reject(error);
|
||
})
|
||
.pipe(destinationStream)
|
||
.on('close', () => {
|
||
resolve();
|
||
})
|
||
.on('error', error => {
|
||
core.error(`An error occurred while writing a downloaded file to ${destinationStream.path}`);
|
||
reject(error);
|
||
});
|
||
}
|
||
else {
|
||
response.message
|
||
.on('error', error => {
|
||
core.error(`An error occurred while attempting to read the response stream`);
|
||
destinationStream.close();
|
||
reject(error);
|
||
})
|
||
.pipe(destinationStream)
|
||
.on('close', () => {
|
||
resolve();
|
||
})
|
||
.on('error', error => {
|
||
core.error(`An error occurred while writing a downloaded file to ${destinationStream.path}`);
|
||
reject(error);
|
||
});
|
||
}
|
||
});
|
||
return;
|
||
});
|
||
}
|
||
}
|
||
exports.DownloadHttpClient = DownloadHttpClient;
|
||
//# sourceMappingURL=download-http-client.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 856:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
module.exports = __webpack_require__(141);
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 862:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __importStar = (this && this.__importStar) || function (mod) {
|
||
if (mod && mod.__esModule) return mod;
|
||
var result = {};
|
||
if (mod != null) for (var k in mod) if (Object.hasOwnProperty.call(mod, k)) result[k] = mod[k];
|
||
result["default"] = mod;
|
||
return result;
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const fs = __importStar(__webpack_require__(747));
|
||
const core_1 = __webpack_require__(470);
|
||
const path_1 = __webpack_require__(622);
|
||
const utils_1 = __webpack_require__(870);
|
||
/**
|
||
* Creates a specification that describes how each file that is part of the artifact will be uploaded
|
||
* @param artifactName the name of the artifact being uploaded. Used during upload to denote where the artifact is stored on the server
|
||
* @param rootDirectory an absolute file path that denotes the path that should be removed from the beginning of each artifact file
|
||
* @param artifactFiles a list of absolute file paths that denote what should be uploaded as part of the artifact
|
||
*/
|
||
function getUploadSpecification(artifactName, rootDirectory, artifactFiles) {
|
||
utils_1.checkArtifactName(artifactName);
|
||
const specifications = [];
|
||
if (!fs.existsSync(rootDirectory)) {
|
||
throw new Error(`Provided rootDirectory ${rootDirectory} does not exist`);
|
||
}
|
||
if (!fs.lstatSync(rootDirectory).isDirectory()) {
|
||
throw new Error(`Provided rootDirectory ${rootDirectory} is not a valid directory`);
|
||
}
|
||
// Normalize and resolve, this allows for either absolute or relative paths to be used
|
||
rootDirectory = path_1.normalize(rootDirectory);
|
||
rootDirectory = path_1.resolve(rootDirectory);
|
||
/*
|
||
Example to demonstrate behavior
|
||
|
||
Input:
|
||
artifactName: my-artifact
|
||
rootDirectory: '/home/user/files/plz-upload'
|
||
artifactFiles: [
|
||
'/home/user/files/plz-upload/file1.txt',
|
||
'/home/user/files/plz-upload/file2.txt',
|
||
'/home/user/files/plz-upload/dir/file3.txt'
|
||
]
|
||
|
||
Output:
|
||
specifications: [
|
||
['/home/user/files/plz-upload/file1.txt', 'my-artifact/file1.txt'],
|
||
['/home/user/files/plz-upload/file1.txt', 'my-artifact/file2.txt'],
|
||
['/home/user/files/plz-upload/file1.txt', 'my-artifact/dir/file3.txt']
|
||
]
|
||
*/
|
||
for (let file of artifactFiles) {
|
||
if (!fs.existsSync(file)) {
|
||
throw new Error(`File ${file} does not exist`);
|
||
}
|
||
if (!fs.lstatSync(file).isDirectory()) {
|
||
// Normalize and resolve, this allows for either absolute or relative paths to be used
|
||
file = path_1.normalize(file);
|
||
file = path_1.resolve(file);
|
||
if (!file.startsWith(rootDirectory)) {
|
||
throw new Error(`The rootDirectory: ${rootDirectory} is not a parent directory of the file: ${file}`);
|
||
}
|
||
// Check for forbidden characters in file paths that will be rejected during upload
|
||
const uploadPath = file.replace(rootDirectory, '');
|
||
utils_1.checkArtifactFilePath(uploadPath);
|
||
/*
|
||
uploadFilePath denotes where the file will be uploaded in the file container on the server. During a run, if multiple artifacts are uploaded, they will all
|
||
be saved in the same container. The artifact name is used as the root directory in the container to separate and distinguish uploaded artifacts
|
||
|
||
path.join handles all the following cases and would return 'artifact-name/file-to-upload.txt
|
||
join('artifact-name/', 'file-to-upload.txt')
|
||
join('artifact-name/', '/file-to-upload.txt')
|
||
join('artifact-name', 'file-to-upload.txt')
|
||
join('artifact-name', '/file-to-upload.txt')
|
||
*/
|
||
specifications.push({
|
||
absoluteFilePath: file,
|
||
uploadFilePath: path_1.join(artifactName, uploadPath)
|
||
});
|
||
}
|
||
else {
|
||
// Directories are rejected by the server during upload
|
||
core_1.debug(`Removing ${file} from rawSearchResults because it is a directory`);
|
||
}
|
||
}
|
||
return specifications;
|
||
}
|
||
exports.getUploadSpecification = getUploadSpecification;
|
||
//# sourceMappingURL=upload-specification.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 864:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var utils = __webpack_require__(35);
|
||
|
||
module.exports = (
|
||
utils.isStandardBrowserEnv() ?
|
||
|
||
// Standard browser envs support document.cookie
|
||
(function standardBrowserEnv() {
|
||
return {
|
||
write: function write(name, value, expires, path, domain, secure) {
|
||
var cookie = [];
|
||
cookie.push(name + '=' + encodeURIComponent(value));
|
||
|
||
if (utils.isNumber(expires)) {
|
||
cookie.push('expires=' + new Date(expires).toGMTString());
|
||
}
|
||
|
||
if (utils.isString(path)) {
|
||
cookie.push('path=' + path);
|
||
}
|
||
|
||
if (utils.isString(domain)) {
|
||
cookie.push('domain=' + domain);
|
||
}
|
||
|
||
if (secure === true) {
|
||
cookie.push('secure');
|
||
}
|
||
|
||
document.cookie = cookie.join('; ');
|
||
},
|
||
|
||
read: function read(name) {
|
||
var match = document.cookie.match(new RegExp('(^|;\\s*)(' + name + ')=([^;]*)'));
|
||
return (match ? decodeURIComponent(match[3]) : null);
|
||
},
|
||
|
||
remove: function remove(name) {
|
||
this.write(name, '', Date.now() - 86400000);
|
||
}
|
||
};
|
||
})() :
|
||
|
||
// Non standard browser env (web workers, react-native) lack needed support.
|
||
(function nonStandardBrowserEnv() {
|
||
return {
|
||
write: function write() {},
|
||
read: function read() { return null; },
|
||
remove: function remove() {}
|
||
};
|
||
})()
|
||
);
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 867:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("tty");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 870:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
||
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
||
return new (P || (P = Promise))(function (resolve, reject) {
|
||
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
||
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
||
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
||
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
||
});
|
||
};
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const core_1 = __webpack_require__(470);
|
||
const fs_1 = __webpack_require__(747);
|
||
const http_client_1 = __webpack_require__(539);
|
||
const auth_1 = __webpack_require__(226);
|
||
const config_variables_1 = __webpack_require__(401);
|
||
/**
|
||
* Returns a retry time in milliseconds that exponentially gets larger
|
||
* depending on the amount of retries that have been attempted
|
||
*/
|
||
function getExponentialRetryTimeInMilliseconds(retryCount) {
|
||
if (retryCount < 0) {
|
||
throw new Error('RetryCount should not be negative');
|
||
}
|
||
else if (retryCount === 0) {
|
||
return config_variables_1.getInitialRetryIntervalInMilliseconds();
|
||
}
|
||
const minTime = config_variables_1.getInitialRetryIntervalInMilliseconds() * config_variables_1.getRetryMultiplier() * retryCount;
|
||
const maxTime = minTime * config_variables_1.getRetryMultiplier();
|
||
// returns a random number between the minTime (inclusive) and the maxTime (exclusive)
|
||
return Math.random() * (maxTime - minTime) + minTime;
|
||
}
|
||
exports.getExponentialRetryTimeInMilliseconds = getExponentialRetryTimeInMilliseconds;
|
||
/**
|
||
* Parses a env variable that is a number
|
||
*/
|
||
function parseEnvNumber(key) {
|
||
const value = Number(process.env[key]);
|
||
if (Number.isNaN(value) || value < 0) {
|
||
return undefined;
|
||
}
|
||
return value;
|
||
}
|
||
exports.parseEnvNumber = parseEnvNumber;
|
||
/**
|
||
* Various utility functions to help with the necessary API calls
|
||
*/
|
||
function getApiVersion() {
|
||
return '6.0-preview';
|
||
}
|
||
exports.getApiVersion = getApiVersion;
|
||
function isSuccessStatusCode(statusCode) {
|
||
if (!statusCode) {
|
||
return false;
|
||
}
|
||
return statusCode >= 200 && statusCode < 300;
|
||
}
|
||
exports.isSuccessStatusCode = isSuccessStatusCode;
|
||
function isForbiddenStatusCode(statusCode) {
|
||
if (!statusCode) {
|
||
return false;
|
||
}
|
||
return statusCode === http_client_1.HttpCodes.Forbidden;
|
||
}
|
||
exports.isForbiddenStatusCode = isForbiddenStatusCode;
|
||
function isRetryableStatusCode(statusCode) {
|
||
if (!statusCode) {
|
||
return false;
|
||
}
|
||
const retryableStatusCodes = [
|
||
http_client_1.HttpCodes.BadGateway,
|
||
http_client_1.HttpCodes.GatewayTimeout,
|
||
http_client_1.HttpCodes.InternalServerError,
|
||
http_client_1.HttpCodes.ServiceUnavailable,
|
||
http_client_1.HttpCodes.TooManyRequests,
|
||
413 // Payload Too Large
|
||
];
|
||
return retryableStatusCodes.includes(statusCode);
|
||
}
|
||
exports.isRetryableStatusCode = isRetryableStatusCode;
|
||
function isThrottledStatusCode(statusCode) {
|
||
if (!statusCode) {
|
||
return false;
|
||
}
|
||
return statusCode === http_client_1.HttpCodes.TooManyRequests;
|
||
}
|
||
exports.isThrottledStatusCode = isThrottledStatusCode;
|
||
/**
|
||
* Attempts to get the retry-after value from a set of http headers. The retry time
|
||
* is originally denoted in seconds, so if present, it is converted to milliseconds
|
||
* @param headers all the headers received when making an http call
|
||
*/
|
||
function tryGetRetryAfterValueTimeInMilliseconds(headers) {
|
||
if (headers['retry-after']) {
|
||
const retryTime = Number(headers['retry-after']);
|
||
if (!isNaN(retryTime)) {
|
||
core_1.info(`Retry-After header is present with a value of ${retryTime}`);
|
||
return retryTime * 1000;
|
||
}
|
||
core_1.info(`Returned retry-after header value: ${retryTime} is non-numeric and cannot be used`);
|
||
return undefined;
|
||
}
|
||
core_1.info(`No retry-after header was found. Dumping all headers for diagnostic purposes`);
|
||
// eslint-disable-next-line no-console
|
||
console.log(headers);
|
||
return undefined;
|
||
}
|
||
exports.tryGetRetryAfterValueTimeInMilliseconds = tryGetRetryAfterValueTimeInMilliseconds;
|
||
function getContentRange(start, end, total) {
|
||
// Format: `bytes start-end/fileSize
|
||
// start and end are inclusive
|
||
// For a 200 byte chunk starting at byte 0:
|
||
// Content-Range: bytes 0-199/200
|
||
return `bytes ${start}-${end}/${total}`;
|
||
}
|
||
exports.getContentRange = getContentRange;
|
||
/**
|
||
* Sets all the necessary headers when downloading an artifact
|
||
* @param {string} contentType the type of content being uploaded
|
||
* @param {boolean} isKeepAlive is the same connection being used to make multiple calls
|
||
* @param {boolean} acceptGzip can we accept a gzip encoded response
|
||
* @param {string} acceptType the type of content that we can accept
|
||
* @returns appropriate headers to make a specific http call during artifact download
|
||
*/
|
||
function getDownloadHeaders(contentType, isKeepAlive, acceptGzip) {
|
||
const requestOptions = {};
|
||
if (contentType) {
|
||
requestOptions['Content-Type'] = contentType;
|
||
}
|
||
if (isKeepAlive) {
|
||
requestOptions['Connection'] = 'Keep-Alive';
|
||
// keep alive for at least 10 seconds before closing the connection
|
||
requestOptions['Keep-Alive'] = '10';
|
||
}
|
||
if (acceptGzip) {
|
||
// if we are expecting a response with gzip encoding, it should be using an octet-stream in the accept header
|
||
requestOptions['Accept-Encoding'] = 'gzip';
|
||
requestOptions['Accept'] = `application/octet-stream;api-version=${getApiVersion()}`;
|
||
}
|
||
else {
|
||
// default to application/json if we are not working with gzip content
|
||
requestOptions['Accept'] = `application/json;api-version=${getApiVersion()}`;
|
||
}
|
||
return requestOptions;
|
||
}
|
||
exports.getDownloadHeaders = getDownloadHeaders;
|
||
/**
|
||
* Sets all the necessary headers when uploading an artifact
|
||
* @param {string} contentType the type of content being uploaded
|
||
* @param {boolean} isKeepAlive is the same connection being used to make multiple calls
|
||
* @param {boolean} isGzip is the connection being used to upload GZip compressed content
|
||
* @param {number} uncompressedLength the original size of the content if something is being uploaded that has been compressed
|
||
* @param {number} contentLength the length of the content that is being uploaded
|
||
* @param {string} contentRange the range of the content that is being uploaded
|
||
* @returns appropriate headers to make a specific http call during artifact upload
|
||
*/
|
||
function getUploadHeaders(contentType, isKeepAlive, isGzip, uncompressedLength, contentLength, contentRange) {
|
||
const requestOptions = {};
|
||
requestOptions['Accept'] = `application/json;api-version=${getApiVersion()}`;
|
||
if (contentType) {
|
||
requestOptions['Content-Type'] = contentType;
|
||
}
|
||
if (isKeepAlive) {
|
||
requestOptions['Connection'] = 'Keep-Alive';
|
||
// keep alive for at least 10 seconds before closing the connection
|
||
requestOptions['Keep-Alive'] = '10';
|
||
}
|
||
if (isGzip) {
|
||
requestOptions['Content-Encoding'] = 'gzip';
|
||
requestOptions['x-tfs-filelength'] = uncompressedLength;
|
||
}
|
||
if (contentLength) {
|
||
requestOptions['Content-Length'] = contentLength;
|
||
}
|
||
if (contentRange) {
|
||
requestOptions['Content-Range'] = contentRange;
|
||
}
|
||
return requestOptions;
|
||
}
|
||
exports.getUploadHeaders = getUploadHeaders;
|
||
function createHttpClient(userAgent) {
|
||
return new http_client_1.HttpClient(userAgent, [
|
||
new auth_1.BearerCredentialHandler(config_variables_1.getRuntimeToken())
|
||
]);
|
||
}
|
||
exports.createHttpClient = createHttpClient;
|
||
function getArtifactUrl() {
|
||
const artifactUrl = `${config_variables_1.getRuntimeUrl()}_apis/pipelines/workflows/${config_variables_1.getWorkFlowRunId()}/artifacts?api-version=${getApiVersion()}`;
|
||
core_1.debug(`Artifact Url: ${artifactUrl}`);
|
||
return artifactUrl;
|
||
}
|
||
exports.getArtifactUrl = getArtifactUrl;
|
||
/**
|
||
* Uh oh! Something might have gone wrong during either upload or download. The IHtttpClientResponse object contains information
|
||
* about the http call that was made by the actions http client. This information might be useful to display for diagnostic purposes, but
|
||
* this entire object is really big and most of the information is not really useful. This function takes the response object and displays only
|
||
* the information that we want.
|
||
*
|
||
* Certain information such as the TLSSocket and the Readable state are not really useful for diagnostic purposes so they can be avoided.
|
||
* Other information such as the headers, the response code and message might be useful, so this is displayed.
|
||
*/
|
||
function displayHttpDiagnostics(response) {
|
||
core_1.info(`##### Begin Diagnostic HTTP information #####
|
||
Status Code: ${response.message.statusCode}
|
||
Status Message: ${response.message.statusMessage}
|
||
Header Information: ${JSON.stringify(response.message.headers, undefined, 2)}
|
||
###### End Diagnostic HTTP information ######`);
|
||
}
|
||
exports.displayHttpDiagnostics = displayHttpDiagnostics;
|
||
/**
|
||
* Invalid characters that cannot be in the artifact name or an uploaded file. Will be rejected
|
||
* from the server if attempted to be sent over. These characters are not allowed due to limitations with certain
|
||
* file systems such as NTFS. To maintain platform-agnostic behavior, all characters that are not supported by an
|
||
* individual filesystem/platform will not be supported on all fileSystems/platforms
|
||
*
|
||
* FilePaths can include characters such as \ and / which are not permitted in the artifact name alone
|
||
*/
|
||
const invalidArtifactFilePathCharacters = ['"', ':', '<', '>', '|', '*', '?'];
|
||
const invalidArtifactNameCharacters = [
|
||
...invalidArtifactFilePathCharacters,
|
||
'\\',
|
||
'/'
|
||
];
|
||
/**
|
||
* Scans the name of the artifact to make sure there are no illegal characters
|
||
*/
|
||
function checkArtifactName(name) {
|
||
if (!name) {
|
||
throw new Error(`Artifact name: ${name}, is incorrectly provided`);
|
||
}
|
||
for (const invalidChar of invalidArtifactNameCharacters) {
|
||
if (name.includes(invalidChar)) {
|
||
throw new Error(`Artifact name is not valid: ${name}. Contains character: "${invalidChar}". Invalid artifact name characters include: ${invalidArtifactNameCharacters.toString()}.`);
|
||
}
|
||
}
|
||
}
|
||
exports.checkArtifactName = checkArtifactName;
|
||
/**
|
||
* Scans the name of the filePath used to make sure there are no illegal characters
|
||
*/
|
||
function checkArtifactFilePath(path) {
|
||
if (!path) {
|
||
throw new Error(`Artifact path: ${path}, is incorrectly provided`);
|
||
}
|
||
for (const invalidChar of invalidArtifactFilePathCharacters) {
|
||
if (path.includes(invalidChar)) {
|
||
throw new Error(`Artifact path is not valid: ${path}. Contains character: "${invalidChar}". Invalid characters include: ${invalidArtifactFilePathCharacters.toString()}.`);
|
||
}
|
||
}
|
||
}
|
||
exports.checkArtifactFilePath = checkArtifactFilePath;
|
||
function createDirectoriesForArtifact(directories) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
for (const directory of directories) {
|
||
yield fs_1.promises.mkdir(directory, {
|
||
recursive: true
|
||
});
|
||
}
|
||
});
|
||
}
|
||
exports.createDirectoriesForArtifact = createDirectoriesForArtifact;
|
||
function createEmptyFilesForArtifact(emptyFilesToCreate) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
for (const filePath of emptyFilesToCreate) {
|
||
yield (yield fs_1.promises.open(filePath, 'w')).close();
|
||
}
|
||
});
|
||
}
|
||
exports.createEmptyFilesForArtifact = createEmptyFilesForArtifact;
|
||
function getFileSize(filePath) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
const stats = yield fs_1.promises.stat(filePath);
|
||
core_1.debug(`${filePath} size:(${stats.size}) blksize:(${stats.blksize}) blocks:(${stats.blocks})`);
|
||
return stats.size;
|
||
});
|
||
}
|
||
exports.getFileSize = getFileSize;
|
||
function rmFile(filePath) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
yield fs_1.promises.unlink(filePath);
|
||
});
|
||
}
|
||
exports.rmFile = rmFile;
|
||
function getProperRetention(retentionInput, retentionSetting) {
|
||
if (retentionInput < 0) {
|
||
throw new Error('Invalid retention, minimum value is 1.');
|
||
}
|
||
let retention = retentionInput;
|
||
if (retentionSetting) {
|
||
const maxRetention = parseInt(retentionSetting);
|
||
if (!isNaN(maxRetention) && maxRetention < retention) {
|
||
core_1.warning(`Retention days is greater than the max value allowed by the repository setting, reduce retention to ${maxRetention} days`);
|
||
retention = maxRetention;
|
||
}
|
||
}
|
||
return retention;
|
||
}
|
||
exports.getProperRetention = getProperRetention;
|
||
function sleep(milliseconds) {
|
||
return __awaiter(this, void 0, void 0, function* () {
|
||
return new Promise(resolve => setTimeout(resolve, milliseconds));
|
||
});
|
||
}
|
||
exports.sleep = sleep;
|
||
//# sourceMappingURL=utils.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 875:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
const {promisify} = __webpack_require__(669);
|
||
const tmp = __webpack_require__(150);
|
||
|
||
// file
|
||
module.exports.fileSync = tmp.fileSync;
|
||
const fileWithOptions = promisify((options, cb) =>
|
||
tmp.file(options, (err, path, fd, cleanup) =>
|
||
err ? cb(err) : cb(undefined, { path, fd, cleanup: promisify(cleanup) })
|
||
)
|
||
);
|
||
module.exports.file = async (options) => fileWithOptions(options);
|
||
|
||
module.exports.withFile = async function withFile(fn, options) {
|
||
const { path, fd, cleanup } = await module.exports.file(options);
|
||
try {
|
||
return await fn({ path, fd });
|
||
} finally {
|
||
await cleanup();
|
||
}
|
||
};
|
||
|
||
|
||
// directory
|
||
module.exports.dirSync = tmp.dirSync;
|
||
const dirWithOptions = promisify((options, cb) =>
|
||
tmp.dir(options, (err, path, cleanup) =>
|
||
err ? cb(err) : cb(undefined, { path, cleanup: promisify(cleanup) })
|
||
)
|
||
);
|
||
module.exports.dir = async (options) => dirWithOptions(options);
|
||
|
||
module.exports.withDir = async function withDir(fn, options) {
|
||
const { path, cleanup } = await module.exports.dir(options);
|
||
try {
|
||
return await fn({ path });
|
||
} finally {
|
||
await cleanup();
|
||
}
|
||
};
|
||
|
||
|
||
// name generation
|
||
module.exports.tmpNameSync = tmp.tmpNameSync;
|
||
module.exports.tmpName = promisify(tmp.tmpName);
|
||
|
||
module.exports.tmpdir = tmp.tmpdir;
|
||
|
||
module.exports.setGracefulCleanup = tmp.setGracefulCleanup;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 879:
|
||
/***/ (function(module) {
|
||
|
||
"use strict";
|
||
|
||
|
||
/**
|
||
* Syntactic sugar for invoking a function and expanding an array for arguments.
|
||
*
|
||
* Common use case would be to use `Function.prototype.apply`.
|
||
*
|
||
* ```js
|
||
* function f(x, y, z) {}
|
||
* var args = [1, 2, 3];
|
||
* f.apply(null, args);
|
||
* ```
|
||
*
|
||
* With `spread` this example can be re-written.
|
||
*
|
||
* ```js
|
||
* spread(function(x, y, z) {})([1, 2, 3]);
|
||
* ```
|
||
*
|
||
* @param {Function} callback
|
||
* @returns {Function}
|
||
*/
|
||
module.exports = function spread(callback) {
|
||
return function wrap(arr) {
|
||
return callback.apply(null, arr);
|
||
};
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 887:
|
||
/***/ (function(module) {
|
||
|
||
"use strict";
|
||
|
||
|
||
/**
|
||
* Creates a new URL by combining the specified URLs
|
||
*
|
||
* @param {string} baseURL The base URL
|
||
* @param {string} relativeURL The relative URL
|
||
* @returns {string} The combined URL
|
||
*/
|
||
module.exports = function combineURLs(baseURL, relativeURL) {
|
||
return relativeURL
|
||
? baseURL.replace(/\/+$/, '') + '/' + relativeURL.replace(/^\/+/, '')
|
||
: baseURL;
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 896:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = function (xs, fn) {
|
||
var res = [];
|
||
for (var i = 0; i < xs.length; i++) {
|
||
var x = fn(xs[i], i);
|
||
if (isArray(x)) res.push.apply(res, x);
|
||
else res.push(x);
|
||
}
|
||
return res;
|
||
};
|
||
|
||
var isArray = Array.isArray || function (xs) {
|
||
return Object.prototype.toString.call(xs) === '[object Array]';
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 903:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("zlib");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 923:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const assert = __webpack_require__(357);
|
||
const os = __webpack_require__(87);
|
||
const path = __webpack_require__(622);
|
||
const pathHelper = __webpack_require__(972);
|
||
const minimatch_1 = __webpack_require__(93);
|
||
const internal_match_kind_1 = __webpack_require__(327);
|
||
const internal_path_1 = __webpack_require__(383);
|
||
const IS_WINDOWS = process.platform === 'win32';
|
||
class Pattern {
|
||
constructor(patternOrNegate, segments) {
|
||
/**
|
||
* Indicates whether matches should be excluded from the result set
|
||
*/
|
||
this.negate = false;
|
||
// Pattern overload
|
||
let pattern;
|
||
if (typeof patternOrNegate === 'string') {
|
||
pattern = patternOrNegate.trim();
|
||
}
|
||
// Segments overload
|
||
else {
|
||
// Convert to pattern
|
||
segments = segments || [];
|
||
assert(segments.length, `Parameter 'segments' must not empty`);
|
||
const root = Pattern.getLiteral(segments[0]);
|
||
assert(root && pathHelper.hasAbsoluteRoot(root), `Parameter 'segments' first element must be a root path`);
|
||
pattern = new internal_path_1.Path(segments).toString().trim();
|
||
if (patternOrNegate) {
|
||
pattern = `!${pattern}`;
|
||
}
|
||
}
|
||
// Negate
|
||
while (pattern.startsWith('!')) {
|
||
this.negate = !this.negate;
|
||
pattern = pattern.substr(1).trim();
|
||
}
|
||
// Normalize slashes and ensures absolute root
|
||
pattern = Pattern.fixupPattern(pattern);
|
||
// Segments
|
||
this.segments = new internal_path_1.Path(pattern).segments;
|
||
// Trailing slash indicates the pattern should only match directories, not regular files
|
||
this.trailingSeparator = pathHelper
|
||
.normalizeSeparators(pattern)
|
||
.endsWith(path.sep);
|
||
pattern = pathHelper.safeTrimTrailingSeparator(pattern);
|
||
// Search path (literal path prior to the first glob segment)
|
||
let foundGlob = false;
|
||
const searchSegments = this.segments
|
||
.map(x => Pattern.getLiteral(x))
|
||
.filter(x => !foundGlob && !(foundGlob = x === ''));
|
||
this.searchPath = new internal_path_1.Path(searchSegments).toString();
|
||
// Root RegExp (required when determining partial match)
|
||
this.rootRegExp = new RegExp(Pattern.regExpEscape(searchSegments[0]), IS_WINDOWS ? 'i' : '');
|
||
// Create minimatch
|
||
const minimatchOptions = {
|
||
dot: true,
|
||
nobrace: true,
|
||
nocase: IS_WINDOWS,
|
||
nocomment: true,
|
||
noext: true,
|
||
nonegate: true
|
||
};
|
||
pattern = IS_WINDOWS ? pattern.replace(/\\/g, '/') : pattern;
|
||
this.minimatch = new minimatch_1.Minimatch(pattern, minimatchOptions);
|
||
}
|
||
/**
|
||
* Matches the pattern against the specified path
|
||
*/
|
||
match(itemPath) {
|
||
// Last segment is globstar?
|
||
if (this.segments[this.segments.length - 1] === '**') {
|
||
// Normalize slashes
|
||
itemPath = pathHelper.normalizeSeparators(itemPath);
|
||
// Append a trailing slash. Otherwise Minimatch will not match the directory immediately
|
||
// preceeding the globstar. For example, given the pattern `/foo/**`, Minimatch returns
|
||
// false for `/foo` but returns true for `/foo/`. Append a trailing slash to handle that quirk.
|
||
if (!itemPath.endsWith(path.sep)) {
|
||
// Note, this is safe because the constructor ensures the pattern has an absolute root.
|
||
// For example, formats like C: and C:foo on Windows are resolved to an aboslute root.
|
||
itemPath = `${itemPath}${path.sep}`;
|
||
}
|
||
}
|
||
else {
|
||
// Normalize slashes and trim unnecessary trailing slash
|
||
itemPath = pathHelper.safeTrimTrailingSeparator(itemPath);
|
||
}
|
||
// Match
|
||
if (this.minimatch.match(itemPath)) {
|
||
return this.trailingSeparator ? internal_match_kind_1.MatchKind.Directory : internal_match_kind_1.MatchKind.All;
|
||
}
|
||
return internal_match_kind_1.MatchKind.None;
|
||
}
|
||
/**
|
||
* Indicates whether the pattern may match descendants of the specified path
|
||
*/
|
||
partialMatch(itemPath) {
|
||
// Normalize slashes and trim unnecessary trailing slash
|
||
itemPath = pathHelper.safeTrimTrailingSeparator(itemPath);
|
||
// matchOne does not handle root path correctly
|
||
if (pathHelper.dirname(itemPath) === itemPath) {
|
||
return this.rootRegExp.test(itemPath);
|
||
}
|
||
return this.minimatch.matchOne(itemPath.split(IS_WINDOWS ? /\\+/ : /\/+/), this.minimatch.set[0], true);
|
||
}
|
||
/**
|
||
* Escapes glob patterns within a path
|
||
*/
|
||
static globEscape(s) {
|
||
return (IS_WINDOWS ? s : s.replace(/\\/g, '\\\\')) // escape '\' on Linux/macOS
|
||
.replace(/(\[)(?=[^/]+\])/g, '[[]') // escape '[' when ']' follows within the path segment
|
||
.replace(/\?/g, '[?]') // escape '?'
|
||
.replace(/\*/g, '[*]'); // escape '*'
|
||
}
|
||
/**
|
||
* Normalizes slashes and ensures absolute root
|
||
*/
|
||
static fixupPattern(pattern) {
|
||
// Empty
|
||
assert(pattern, 'pattern cannot be empty');
|
||
// Must not contain `.` segment, unless first segment
|
||
// Must not contain `..` segment
|
||
const literalSegments = new internal_path_1.Path(pattern).segments.map(x => Pattern.getLiteral(x));
|
||
assert(literalSegments.every((x, i) => (x !== '.' || i === 0) && x !== '..'), `Invalid pattern '${pattern}'. Relative pathing '.' and '..' is not allowed.`);
|
||
// Must not contain globs in root, e.g. Windows UNC path \\foo\b*r
|
||
assert(!pathHelper.hasRoot(pattern) || literalSegments[0], `Invalid pattern '${pattern}'. Root segment must not contain globs.`);
|
||
// Normalize slashes
|
||
pattern = pathHelper.normalizeSeparators(pattern);
|
||
// Replace leading `.` segment
|
||
if (pattern === '.' || pattern.startsWith(`.${path.sep}`)) {
|
||
pattern = Pattern.globEscape(process.cwd()) + pattern.substr(1);
|
||
}
|
||
// Replace leading `~` segment
|
||
else if (pattern === '~' || pattern.startsWith(`~${path.sep}`)) {
|
||
const homedir = os.homedir();
|
||
assert(homedir, 'Unable to determine HOME directory');
|
||
assert(pathHelper.hasAbsoluteRoot(homedir), `Expected HOME directory to be a rooted path. Actual '${homedir}'`);
|
||
pattern = Pattern.globEscape(homedir) + pattern.substr(1);
|
||
}
|
||
// Replace relative drive root, e.g. pattern is C: or C:foo
|
||
else if (IS_WINDOWS &&
|
||
(pattern.match(/^[A-Z]:$/i) || pattern.match(/^[A-Z]:[^\\]/i))) {
|
||
let root = pathHelper.ensureAbsoluteRoot('C:\\dummy-root', pattern.substr(0, 2));
|
||
if (pattern.length > 2 && !root.endsWith('\\')) {
|
||
root += '\\';
|
||
}
|
||
pattern = Pattern.globEscape(root) + pattern.substr(2);
|
||
}
|
||
// Replace relative root, e.g. pattern is \ or \foo
|
||
else if (IS_WINDOWS && (pattern === '\\' || pattern.match(/^\\[^\\]/))) {
|
||
let root = pathHelper.ensureAbsoluteRoot('C:\\dummy-root', '\\');
|
||
if (!root.endsWith('\\')) {
|
||
root += '\\';
|
||
}
|
||
pattern = Pattern.globEscape(root) + pattern.substr(1);
|
||
}
|
||
// Otherwise ensure absolute root
|
||
else {
|
||
pattern = pathHelper.ensureAbsoluteRoot(Pattern.globEscape(process.cwd()), pattern);
|
||
}
|
||
return pathHelper.normalizeSeparators(pattern);
|
||
}
|
||
/**
|
||
* Attempts to unescape a pattern segment to create a literal path segment.
|
||
* Otherwise returns empty string.
|
||
*/
|
||
static getLiteral(segment) {
|
||
let literal = '';
|
||
for (let i = 0; i < segment.length; i++) {
|
||
const c = segment[i];
|
||
// Escape
|
||
if (c === '\\' && !IS_WINDOWS && i + 1 < segment.length) {
|
||
literal += segment[++i];
|
||
continue;
|
||
}
|
||
// Wildcard
|
||
else if (c === '*' || c === '?') {
|
||
return '';
|
||
}
|
||
// Character set
|
||
else if (c === '[' && i + 1 < segment.length) {
|
||
let set = '';
|
||
let closed = -1;
|
||
for (let i2 = i + 1; i2 < segment.length; i2++) {
|
||
const c2 = segment[i2];
|
||
// Escape
|
||
if (c2 === '\\' && !IS_WINDOWS && i2 + 1 < segment.length) {
|
||
set += segment[++i2];
|
||
continue;
|
||
}
|
||
// Closed
|
||
else if (c2 === ']') {
|
||
closed = i2;
|
||
break;
|
||
}
|
||
// Otherwise
|
||
else {
|
||
set += c2;
|
||
}
|
||
}
|
||
// Closed?
|
||
if (closed >= 0) {
|
||
// Cannot convert
|
||
if (set.length > 1) {
|
||
return '';
|
||
}
|
||
// Convert to literal
|
||
if (set) {
|
||
literal += set;
|
||
i = closed;
|
||
continue;
|
||
}
|
||
}
|
||
// Otherwise fall thru
|
||
}
|
||
// Append
|
||
literal += c;
|
||
}
|
||
return literal;
|
||
}
|
||
/**
|
||
* Escapes regexp special characters
|
||
* https://javascript.info/regexp-escaping
|
||
*/
|
||
static regExpEscape(s) {
|
||
return s.replace(/[[\\^$.|?*+()]/g, '\\$&');
|
||
}
|
||
}
|
||
exports.Pattern = Pattern;
|
||
//# sourceMappingURL=internal-pattern.js.map
|
||
|
||
/***/ }),
|
||
|
||
/***/ 937:
|
||
/***/ (function(module) {
|
||
|
||
module.exports = require("net");
|
||
|
||
/***/ }),
|
||
|
||
/***/ 946:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var utils = __webpack_require__(35);
|
||
var transformData = __webpack_require__(589);
|
||
var isCancel = __webpack_require__(732);
|
||
var defaults = __webpack_require__(529);
|
||
|
||
/**
|
||
* Throws a `Cancel` if cancellation has been requested.
|
||
*/
|
||
function throwIfCancellationRequested(config) {
|
||
if (config.cancelToken) {
|
||
config.cancelToken.throwIfRequested();
|
||
}
|
||
}
|
||
|
||
/**
|
||
* Dispatch a request to the server using the configured adapter.
|
||
*
|
||
* @param {object} config The config that is to be used for the request
|
||
* @returns {Promise} The Promise to be fulfilled
|
||
*/
|
||
module.exports = function dispatchRequest(config) {
|
||
throwIfCancellationRequested(config);
|
||
|
||
// Ensure headers exist
|
||
config.headers = config.headers || {};
|
||
|
||
// Transform request data
|
||
config.data = transformData.call(
|
||
config,
|
||
config.data,
|
||
config.headers,
|
||
config.transformRequest
|
||
);
|
||
|
||
// Flatten headers
|
||
config.headers = utils.merge(
|
||
config.headers.common || {},
|
||
config.headers[config.method] || {},
|
||
config.headers
|
||
);
|
||
|
||
utils.forEach(
|
||
['delete', 'get', 'head', 'post', 'put', 'patch', 'common'],
|
||
function cleanHeaderConfig(method) {
|
||
delete config.headers[method];
|
||
}
|
||
);
|
||
|
||
var adapter = config.adapter || defaults.adapter;
|
||
|
||
return adapter(config).then(function onAdapterResolution(response) {
|
||
throwIfCancellationRequested(config);
|
||
|
||
// Transform response data
|
||
response.data = transformData.call(
|
||
config,
|
||
response.data,
|
||
response.headers,
|
||
config.transformResponse
|
||
);
|
||
|
||
return response;
|
||
}, function onAdapterRejection(reason) {
|
||
if (!isCancel(reason)) {
|
||
throwIfCancellationRequested(config);
|
||
|
||
// Transform response data
|
||
if (reason && reason.response) {
|
||
reason.response.data = transformData.call(
|
||
config,
|
||
reason.response.data,
|
||
reason.response.headers,
|
||
config.transformResponse
|
||
);
|
||
}
|
||
}
|
||
|
||
return Promise.reject(reason);
|
||
});
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 950:
|
||
/***/ (function(__unusedmodule, exports) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
function getProxyUrl(reqUrl) {
|
||
let usingSsl = reqUrl.protocol === 'https:';
|
||
let proxyUrl;
|
||
if (checkBypass(reqUrl)) {
|
||
return proxyUrl;
|
||
}
|
||
let proxyVar;
|
||
if (usingSsl) {
|
||
proxyVar = process.env['https_proxy'] || process.env['HTTPS_PROXY'];
|
||
}
|
||
else {
|
||
proxyVar = process.env['http_proxy'] || process.env['HTTP_PROXY'];
|
||
}
|
||
if (proxyVar) {
|
||
proxyUrl = new URL(proxyVar);
|
||
}
|
||
return proxyUrl;
|
||
}
|
||
exports.getProxyUrl = getProxyUrl;
|
||
function checkBypass(reqUrl) {
|
||
if (!reqUrl.hostname) {
|
||
return false;
|
||
}
|
||
let noProxy = process.env['no_proxy'] || process.env['NO_PROXY'] || '';
|
||
if (!noProxy) {
|
||
return false;
|
||
}
|
||
// Determine the request port
|
||
let reqPort;
|
||
if (reqUrl.port) {
|
||
reqPort = Number(reqUrl.port);
|
||
}
|
||
else if (reqUrl.protocol === 'http:') {
|
||
reqPort = 80;
|
||
}
|
||
else if (reqUrl.protocol === 'https:') {
|
||
reqPort = 443;
|
||
}
|
||
// Format the request hostname and hostname with port
|
||
let upperReqHosts = [reqUrl.hostname.toUpperCase()];
|
||
if (typeof reqPort === 'number') {
|
||
upperReqHosts.push(`${upperReqHosts[0]}:${reqPort}`);
|
||
}
|
||
// Compare request host against noproxy
|
||
for (let upperNoProxyItem of noProxy
|
||
.split(',')
|
||
.map(x => x.trim().toUpperCase())
|
||
.filter(x => x)) {
|
||
if (upperReqHosts.some(x => x === upperNoProxyItem)) {
|
||
return true;
|
||
}
|
||
}
|
||
return false;
|
||
}
|
||
exports.checkBypass = checkBypass;
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 960:
|
||
/***/ (function(module, __unusedexports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
|
||
var isAbsoluteURL = __webpack_require__(590);
|
||
var combineURLs = __webpack_require__(887);
|
||
|
||
/**
|
||
* Creates a new URL by combining the baseURL with the requestedURL,
|
||
* only when the requestedURL is not already an absolute URL.
|
||
* If the requestURL is absolute, this function returns the requestedURL untouched.
|
||
*
|
||
* @param {string} baseURL The base URL
|
||
* @param {string} requestedURL Absolute or relative URL to combine
|
||
* @returns {string} The combined full path
|
||
*/
|
||
module.exports = function buildFullPath(baseURL, requestedURL) {
|
||
if (baseURL && !isAbsoluteURL(requestedURL)) {
|
||
return combineURLs(baseURL, requestedURL);
|
||
}
|
||
return requestedURL;
|
||
};
|
||
|
||
|
||
/***/ }),
|
||
|
||
/***/ 972:
|
||
/***/ (function(__unusedmodule, exports, __webpack_require__) {
|
||
|
||
"use strict";
|
||
|
||
Object.defineProperty(exports, "__esModule", { value: true });
|
||
const assert = __webpack_require__(357);
|
||
const path = __webpack_require__(622);
|
||
const IS_WINDOWS = process.platform === 'win32';
|
||
/**
|
||
* Similar to path.dirname except normalizes the path separators and slightly better handling for Windows UNC paths.
|
||
*
|
||
* For example, on Linux/macOS:
|
||
* - `/ => /`
|
||
* - `/hello => /`
|
||
*
|
||
* For example, on Windows:
|
||
* - `C:\ => C:\`
|
||
* - `C:\hello => C:\`
|
||
* - `C: => C:`
|
||
* - `C:hello => C:`
|
||
* - `\ => \`
|
||
* - `\hello => \`
|
||
* - `\\hello => \\hello`
|
||
* - `\\hello\world => \\hello\world`
|
||
*/
|
||
function dirname(p) {
|
||
// Normalize slashes and trim unnecessary trailing slash
|
||
p = safeTrimTrailingSeparator(p);
|
||
// Windows UNC root, e.g. \\hello or \\hello\world
|
||
if (IS_WINDOWS && /^\\\\[^\\]+(\\[^\\]+)?$/.test(p)) {
|
||
return p;
|
||
}
|
||
// Get dirname
|
||
let result = path.dirname(p);
|
||
// Trim trailing slash for Windows UNC root, e.g. \\hello\world\
|
||
if (IS_WINDOWS && /^\\\\[^\\]+\\[^\\]+\\$/.test(result)) {
|
||
result = safeTrimTrailingSeparator(result);
|
||
}
|
||
return result;
|
||
}
|
||
exports.dirname = dirname;
|
||
/**
|
||
* Roots the path if not already rooted. On Windows, relative roots like `\`
|
||
* or `C:` are expanded based on the current working directory.
|
||
*/
|
||
function ensureAbsoluteRoot(root, itemPath) {
|
||
assert(root, `ensureAbsoluteRoot parameter 'root' must not be empty`);
|
||
assert(itemPath, `ensureAbsoluteRoot parameter 'itemPath' must not be empty`);
|
||
// Already rooted
|
||
if (hasAbsoluteRoot(itemPath)) {
|
||
return itemPath;
|
||
}
|
||
// Windows
|
||
if (IS_WINDOWS) {
|
||
// Check for itemPath like C: or C:foo
|
||
if (itemPath.match(/^[A-Z]:[^\\/]|^[A-Z]:$/i)) {
|
||
let cwd = process.cwd();
|
||
assert(cwd.match(/^[A-Z]:\\/i), `Expected current directory to start with an absolute drive root. Actual '${cwd}'`);
|
||
// Drive letter matches cwd? Expand to cwd
|
||
if (itemPath[0].toUpperCase() === cwd[0].toUpperCase()) {
|
||
// Drive only, e.g. C:
|
||
if (itemPath.length === 2) {
|
||
// Preserve specified drive letter case (upper or lower)
|
||
return `${itemPath[0]}:\\${cwd.substr(3)}`;
|
||
}
|
||
// Drive + path, e.g. C:foo
|
||
else {
|
||
if (!cwd.endsWith('\\')) {
|
||
cwd += '\\';
|
||
}
|
||
// Preserve specified drive letter case (upper or lower)
|
||
return `${itemPath[0]}:\\${cwd.substr(3)}${itemPath.substr(2)}`;
|
||
}
|
||
}
|
||
// Different drive
|
||
else {
|
||
return `${itemPath[0]}:\\${itemPath.substr(2)}`;
|
||
}
|
||
}
|
||
// Check for itemPath like \ or \foo
|
||
else if (normalizeSeparators(itemPath).match(/^\\$|^\\[^\\]/)) {
|
||
const cwd = process.cwd();
|
||
assert(cwd.match(/^[A-Z]:\\/i), `Expected current directory to start with an absolute drive root. Actual '${cwd}'`);
|
||
return `${cwd[0]}:\\${itemPath.substr(1)}`;
|
||
}
|
||
}
|
||
assert(hasAbsoluteRoot(root), `ensureAbsoluteRoot parameter 'root' must have an absolute root`);
|
||
// Otherwise ensure root ends with a separator
|
||
if (root.endsWith('/') || (IS_WINDOWS && root.endsWith('\\'))) {
|
||
// Intentionally empty
|
||
}
|
||
else {
|
||
// Append separator
|
||
root += path.sep;
|
||
}
|
||
return root + itemPath;
|
||
}
|
||
exports.ensureAbsoluteRoot = ensureAbsoluteRoot;
|
||
/**
|
||
* On Linux/macOS, true if path starts with `/`. On Windows, true for paths like:
|
||
* `\\hello\share` and `C:\hello` (and using alternate separator).
|
||
*/
|
||
function hasAbsoluteRoot(itemPath) {
|
||
assert(itemPath, `hasAbsoluteRoot parameter 'itemPath' must not be empty`);
|
||
// Normalize separators
|
||
itemPath = normalizeSeparators(itemPath);
|
||
// Windows
|
||
if (IS_WINDOWS) {
|
||
// E.g. \\hello\share or C:\hello
|
||
return itemPath.startsWith('\\\\') || /^[A-Z]:\\/i.test(itemPath);
|
||
}
|
||
// E.g. /hello
|
||
return itemPath.startsWith('/');
|
||
}
|
||
exports.hasAbsoluteRoot = hasAbsoluteRoot;
|
||
/**
|
||
* On Linux/macOS, true if path starts with `/`. On Windows, true for paths like:
|
||
* `\`, `\hello`, `\\hello\share`, `C:`, and `C:\hello` (and using alternate separator).
|
||
*/
|
||
function hasRoot(itemPath) {
|
||
assert(itemPath, `isRooted parameter 'itemPath' must not be empty`);
|
||
// Normalize separators
|
||
itemPath = normalizeSeparators(itemPath);
|
||
// Windows
|
||
if (IS_WINDOWS) {
|
||
// E.g. \ or \hello or \\hello
|
||
// E.g. C: or C:\hello
|
||
return itemPath.startsWith('\\') || /^[A-Z]:/i.test(itemPath);
|
||
}
|
||
// E.g. /hello
|
||
return itemPath.startsWith('/');
|
||
}
|
||
exports.hasRoot = hasRoot;
|
||
/**
|
||
* Removes redundant slashes and converts `/` to `\` on Windows
|
||
*/
|
||
function normalizeSeparators(p) {
|
||
p = p || '';
|
||
// Windows
|
||
if (IS_WINDOWS) {
|
||
// Convert slashes on Windows
|
||
p = p.replace(/\//g, '\\');
|
||
// Remove redundant slashes
|
||
const isUnc = /^\\\\+[^\\]/.test(p); // e.g. \\hello
|
||
return (isUnc ? '\\' : '') + p.replace(/\\\\+/g, '\\'); // preserve leading \\ for UNC
|
||
}
|
||
// Remove redundant slashes
|
||
return p.replace(/\/\/+/g, '/');
|
||
}
|
||
exports.normalizeSeparators = normalizeSeparators;
|
||
/**
|
||
* Normalizes the path separators and trims the trailing separator (when safe).
|
||
* For example, `/foo/ => /foo` but `/ => /`
|
||
*/
|
||
function safeTrimTrailingSeparator(p) {
|
||
// Short-circuit if empty
|
||
if (!p) {
|
||
return '';
|
||
}
|
||
// Normalize separators
|
||
p = normalizeSeparators(p);
|
||
// No trailing slash
|
||
if (!p.endsWith(path.sep)) {
|
||
return p;
|
||
}
|
||
// Check '/' on Linux/macOS and '\' on Windows
|
||
if (p === path.sep) {
|
||
return p;
|
||
}
|
||
// On Windows check if drive root. E.g. C:\
|
||
if (IS_WINDOWS && /^[A-Z]:\\$/i.test(p)) {
|
||
return p;
|
||
}
|
||
// Otherwise trim trailing slash
|
||
return p.substr(0, p.length - 1);
|
||
}
|
||
exports.safeTrimTrailingSeparator = safeTrimTrailingSeparator;
|
||
//# sourceMappingURL=internal-path-helper.js.map
|
||
|
||
/***/ })
|
||
|
||
/******/ }); |