From 5091e42a03ad8798b48c3e7f7a52c0055dba9e34 Mon Sep 17 00:00:00 2001 From: Kasper Svendsen Date: Thu, 13 Nov 2025 10:44:19 +0100 Subject: [PATCH 01/51] Overlay: Remove repository owner restriction --- lib/init-action.js | 9 ++------- src/config-utils.test.ts | 43 ++-------------------------------------- src/config-utils.ts | 9 --------- 3 files changed, 4 insertions(+), 57 deletions(-) diff --git a/lib/init-action.js b/lib/init-action.js index 98c23c88f..7ec583c51 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -86884,10 +86884,7 @@ var OVERLAY_ANALYSIS_CODE_SCANNING_FEATURES = { rust: "overlay_analysis_code_scanning_rust" /* OverlayAnalysisCodeScanningRust */, swift: "overlay_analysis_code_scanning_swift" /* OverlayAnalysisCodeScanningSwift */ }; -async function isOverlayAnalysisFeatureEnabled(repository, features, codeql, languages, codeScanningConfig) { - if (!["github", "dsp-testing"].includes(repository.owner)) { - return false; - } +async function isOverlayAnalysisFeatureEnabled(features, codeql, languages, codeScanningConfig) { if (!await features.getValue("overlay_analysis" /* OverlayAnalysis */, codeql)) { return false; } @@ -86909,7 +86906,7 @@ async function isOverlayAnalysisFeatureEnabled(repository, features, codeql, lan } return true; } -async function getOverlayDatabaseMode(codeql, repository, features, languages, sourceRoot, buildMode, codeScanningConfig, logger) { +async function getOverlayDatabaseMode(codeql, features, languages, sourceRoot, buildMode, codeScanningConfig, logger) { let overlayDatabaseMode = "none" /* None */; let useOverlayDatabaseCaching = false; const modeEnv = process.env.CODEQL_OVERLAY_DATABASE_MODE; @@ -86919,7 +86916,6 @@ async function getOverlayDatabaseMode(codeql, repository, features, languages, s `Setting overlay database mode to ${overlayDatabaseMode} from the CODEQL_OVERLAY_DATABASE_MODE environment variable.` ); } else if (await isOverlayAnalysisFeatureEnabled( - repository, features, codeql, languages, @@ -87027,7 +87023,6 @@ async function initConfig(features, inputs) { } const { overlayDatabaseMode, useOverlayDatabaseCaching } = await getOverlayDatabaseMode( inputs.codeql, - inputs.repository, inputs.features, config.languages, inputs.sourceRoot, diff --git a/src/config-utils.test.ts b/src/config-utils.test.ts index 32c794450..da58dd8b1 100644 --- a/src/config-utils.test.ts +++ b/src/config-utils.test.ts @@ -990,7 +990,6 @@ interface OverlayDatabaseModeTestSetup { features: Feature[]; isPullRequest: boolean; isDefaultBranch: boolean; - repositoryOwner: string; buildMode: BuildMode | undefined; languages: Language[]; codeqlVersion: string; @@ -1003,7 +1002,6 @@ const defaultOverlayDatabaseModeTestSetup: OverlayDatabaseModeTestSetup = { features: [], isPullRequest: false, isDefaultBranch: false, - repositoryOwner: "github", buildMode: BuildMode.None, languages: [KnownLanguage.javascript], codeqlVersion: CODEQL_OVERLAY_MINIMUM_VERSION, @@ -1049,12 +1047,6 @@ const getOverlayDatabaseModeMacro = test.macro({ .stub(actionsUtil, "isAnalyzingPullRequest") .returns(setup.isPullRequest); - // Mock repository owner - const repository = { - owner: setup.repositoryOwner, - repo: "test-repo", - }; - // Set up CodeQL mock const codeql = mockCodeQLVersion(setup.codeqlVersion); @@ -1077,7 +1069,6 @@ const getOverlayDatabaseModeMacro = test.macro({ const result = await configUtils.getOverlayDatabaseMode( codeql, - repository, features, setup.languages, tempDir, // sourceRoot @@ -1499,10 +1490,9 @@ test( test( getOverlayDatabaseModeMacro, - "Overlay PR analysis by env for dsp-testing", + "Overlay PR analysis by env", { overlayDatabaseEnvVar: "overlay", - repositoryOwner: "dsp-testing", }, { overlayDatabaseMode: OverlayDatabaseMode.Overlay, @@ -1512,25 +1502,11 @@ test( test( getOverlayDatabaseModeMacro, - "Overlay PR analysis by env for other-org", - { - overlayDatabaseEnvVar: "overlay", - repositoryOwner: "other-org", - }, - { - overlayDatabaseMode: OverlayDatabaseMode.Overlay, - useOverlayDatabaseCaching: false, - }, -); - -test( - getOverlayDatabaseModeMacro, - "Overlay PR analysis by feature flag for dsp-testing", + "Overlay PR analysis by feature flag", { languages: [KnownLanguage.javascript], features: [Feature.OverlayAnalysis, Feature.OverlayAnalysisJavascript], isPullRequest: true, - repositoryOwner: "dsp-testing", }, { overlayDatabaseMode: OverlayDatabaseMode.Overlay, @@ -1538,21 +1514,6 @@ test( }, ); -test( - getOverlayDatabaseModeMacro, - "No overlay PR analysis by feature flag for other-org", - { - languages: [KnownLanguage.javascript], - features: [Feature.OverlayAnalysis, Feature.OverlayAnalysisJavascript], - isPullRequest: true, - repositoryOwner: "other-org", - }, - { - overlayDatabaseMode: OverlayDatabaseMode.None, - useOverlayDatabaseCaching: false, - }, -); - test( getOverlayDatabaseModeMacro, "Fallback due to autobuild with traced language", diff --git a/src/config-utils.ts b/src/config-utils.ts index 0f152a633..fa88d4b4a 100644 --- a/src/config-utils.ts +++ b/src/config-utils.ts @@ -579,17 +579,11 @@ const OVERLAY_ANALYSIS_CODE_SCANNING_FEATURES: Record = { }; async function isOverlayAnalysisFeatureEnabled( - repository: RepositoryNwo, features: FeatureEnablement, codeql: CodeQL, languages: Language[], codeScanningConfig: UserConfig, ): Promise { - // TODO: Remove the repository owner check once support for overlay analysis - // stabilizes, and no more backward-incompatible changes are expected. - if (!["github", "dsp-testing"].includes(repository.owner)) { - return false; - } if (!(await features.getValue(Feature.OverlayAnalysis, codeql))) { return false; } @@ -647,7 +641,6 @@ async function isOverlayAnalysisFeatureEnabled( */ export async function getOverlayDatabaseMode( codeql: CodeQL, - repository: RepositoryNwo, features: FeatureEnablement, languages: Language[], sourceRoot: string, @@ -676,7 +669,6 @@ export async function getOverlayDatabaseMode( ); } else if ( await isOverlayAnalysisFeatureEnabled( - repository, features, codeql, languages, @@ -846,7 +838,6 @@ export async function initConfig( const { overlayDatabaseMode, useOverlayDatabaseCaching } = await getOverlayDatabaseMode( inputs.codeql, - inputs.repository, inputs.features, config.languages, inputs.sourceRoot, From 362f8d1d2dd58d870a49b48980299570b8daa574 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Thu, 13 Nov 2025 18:39:52 +0000 Subject: [PATCH 02/51] Update default bundle to codeql-bundle-v2.23.5 --- lib/analyze-action.js | 4 ++-- lib/autobuild-action.js | 4 ++-- lib/defaults.json | 8 ++++---- lib/init-action-post.js | 4 ++-- lib/init-action.js | 4 ++-- lib/setup-codeql-action.js | 4 ++-- lib/start-proxy-action.js | 4 ++-- lib/upload-lib.js | 4 ++-- lib/upload-sarif-action.js | 4 ++-- src/defaults.json | 8 ++++---- 10 files changed, 24 insertions(+), 24 deletions(-) diff --git a/lib/analyze-action.js b/lib/analyze-action.js index fd43a2f77..deb4af906 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -88210,8 +88210,8 @@ var path4 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.5"; +var cliVersion = "2.23.5"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index d09fe07d8..fdeb4d70d 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -83700,8 +83700,8 @@ var path3 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.5"; +var cliVersion = "2.23.5"; // src/overlay-database-utils.ts var fs2 = __toESM(require("fs")); diff --git a/lib/defaults.json b/lib/defaults.json index 1fa392f54..9be5b5476 100644 --- a/lib/defaults.json +++ b/lib/defaults.json @@ -1,6 +1,6 @@ { - "bundleVersion": "codeql-bundle-v2.23.3", - "cliVersion": "2.23.3", - "priorBundleVersion": "codeql-bundle-v2.23.2", - "priorCliVersion": "2.23.2" + "bundleVersion": "codeql-bundle-v2.23.5", + "cliVersion": "2.23.5", + "priorBundleVersion": "codeql-bundle-v2.23.3", + "priorCliVersion": "2.23.3" } diff --git a/lib/init-action-post.js b/lib/init-action-post.js index 1d3c4d5d9..ef36db529 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -122974,8 +122974,8 @@ var path4 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.5"; +var cliVersion = "2.23.5"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); diff --git a/lib/init-action.js b/lib/init-action.js index 98c23c88f..41cdafba1 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -85634,8 +85634,8 @@ var path5 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.5"; +var cliVersion = "2.23.5"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index c9e95730b..e119b9477 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -83588,8 +83588,8 @@ var path4 = __toESM(require("path")); var semver3 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.5"; +var cliVersion = "2.23.5"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index 281341e5a..2af00a59a 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -99683,8 +99683,8 @@ function getActionsLogger() { var core7 = __toESM(require_core()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.5"; +var cliVersion = "2.23.5"; // src/languages.ts var KnownLanguage = /* @__PURE__ */ ((KnownLanguage2) => { diff --git a/lib/upload-lib.js b/lib/upload-lib.js index 2e980ba46..dc7d0d162 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -86724,8 +86724,8 @@ var path4 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.5"; +var cliVersion = "2.23.5"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 6fd196c32..46c84e88c 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -86504,8 +86504,8 @@ var path4 = __toESM(require("path")); var semver3 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.5"; +var cliVersion = "2.23.5"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); diff --git a/src/defaults.json b/src/defaults.json index 1fa392f54..9be5b5476 100644 --- a/src/defaults.json +++ b/src/defaults.json @@ -1,6 +1,6 @@ { - "bundleVersion": "codeql-bundle-v2.23.3", - "cliVersion": "2.23.3", - "priorBundleVersion": "codeql-bundle-v2.23.2", - "priorCliVersion": "2.23.2" + "bundleVersion": "codeql-bundle-v2.23.5", + "cliVersion": "2.23.5", + "priorBundleVersion": "codeql-bundle-v2.23.3", + "priorCliVersion": "2.23.3" } From 8d3d4001e38901ebbee39c134163b67198c956f7 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Thu, 13 Nov 2025 18:40:00 +0000 Subject: [PATCH 03/51] Add changelog note --- CHANGELOG.md | 1 + 1 file changed, 1 insertion(+) diff --git a/CHANGELOG.md b/CHANGELOG.md index b023f376b..923cc8264 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -5,6 +5,7 @@ See the [releases page](https://github.com/github/codeql-action/releases) for th ## [UNRELEASED] - CodeQL Action v3 will be deprecated in December 2026. The Action now logs a warning for customers who are running v3 but could be running v4. For more information, see [Upcoming deprecation of CodeQL Action v3](https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/). +- Update default CodeQL bundle version to 2.23.5. [#3288](https://github.com/github/codeql-action/pull/3288) ## 4.31.2 - 30 Oct 2025 From f20e02164a8bd2f32913932752d67ee2bbf22246 Mon Sep 17 00:00:00 2001 From: "Michael B. Gale" Date: Thu, 13 Nov 2025 18:58:54 +0000 Subject: [PATCH 04/51] Add support for adding `setup-dotnet` steps to `sync.sh` --- .github/workflows/__local-bundle.yml | 14 ++++++++++++++ pr-checks/checks/local-bundle.yml | 1 + pr-checks/sync.py | 19 +++++++++++++++++++ 3 files changed, 34 insertions(+) diff --git a/.github/workflows/__local-bundle.yml b/.github/workflows/__local-bundle.yml index 159d8c455..3fc89f381 100644 --- a/.github/workflows/__local-bundle.yml +++ b/.github/workflows/__local-bundle.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -85,6 +95,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - name: Fetch latest CodeQL bundle run: | wget https://github.com/github/codeql-action/releases/latest/download/codeql-bundle-linux64.tar.zst diff --git a/pr-checks/checks/local-bundle.yml b/pr-checks/checks/local-bundle.yml index 449dcbbb1..c0930772e 100644 --- a/pr-checks/checks/local-bundle.yml +++ b/pr-checks/checks/local-bundle.yml @@ -3,6 +3,7 @@ description: "Tests using a CodeQL bundle from a local file rather than a URL" versions: ["linked"] installGo: true installPython: true +installDotNet: true steps: - name: Fetch latest CodeQL bundle run: | diff --git a/pr-checks/sync.py b/pr-checks/sync.py index f247f3824..77816be76 100755 --- a/pr-checks/sync.py +++ b/pr-checks/sync.py @@ -204,6 +204,25 @@ for file in sorted((this_dir / 'checks').glob('*.yml')): } }) + installDotNet = is_truthy(checkSpecification.get('installDotNet', '')) + + if installDotNet: + baseDotNetVersionExpr = '9.x' + workflowInputs['dotnet-version'] = { + 'type': 'string', + 'description': 'The version of .NET to install', + 'required': False, + 'default': baseDotNetVersionExpr, + } + + steps.append({ + 'name': 'Install .NET', + 'uses': 'actions/setup-dotnet@v5', + 'with': { + 'dotnet-version': '${{ inputs.dotnet-version || \'' + baseDotNetVersionExpr + '\' }}' + } + }) + # If container initialisation steps are present in the check specification, # make sure to execute them first. if 'container' in checkSpecification and 'container-init-steps' in checkSpecification: From 58c9eb6c034b7054387301aa21926d94da049b69 Mon Sep 17 00:00:00 2001 From: "Michael B. Gale" Date: Thu, 13 Nov 2025 19:17:53 +0000 Subject: [PATCH 05/51] Add `global.json` --- tests/multi-language-repo/global.json | 6 ++++++ 1 file changed, 6 insertions(+) create mode 100644 tests/multi-language-repo/global.json diff --git a/tests/multi-language-repo/global.json b/tests/multi-language-repo/global.json new file mode 100644 index 000000000..764fdd82b --- /dev/null +++ b/tests/multi-language-repo/global.json @@ -0,0 +1,6 @@ +{ + "sdk": { + "version": "9.0.307", + "rollForward": "latestFeature" + } +} From 38a3a7258f252b705a070bd91df7a72a50d61318 Mon Sep 17 00:00:00 2001 From: "Michael B. Gale" Date: Thu, 13 Nov 2025 19:08:57 +0000 Subject: [PATCH 06/51] Enable `installDotNet` in all workflows that analyse C# --- .github/workflows/__all-platform-bundle.yml | 14 ++++++++++++++ .github/workflows/__analyze-ref-input.yml | 14 ++++++++++++++ .github/workflows/__autobuild-action.yml | 18 ++++++++++++++++-- .github/workflows/__build-mode-manual.yml | 14 ++++++++++++++ .../__export-file-baseline-information.yml | 14 ++++++++++++++ .github/workflows/__go-custom-queries.yml | 14 ++++++++++++++ .github/workflows/__go.yml | 6 ++++++ .../workflows/__multi-language-autodetect.yml | 14 ++++++++++++++ ...packaging-codescanning-config-inputs-js.yml | 14 ++++++++++++++ .../workflows/__packaging-config-inputs-js.yml | 14 ++++++++++++++ .github/workflows/__packaging-config-js.yml | 14 ++++++++++++++ .github/workflows/__packaging-inputs-js.yml | 14 ++++++++++++++ .github/workflows/__remote-config.yml | 14 ++++++++++++++ .github/workflows/__split-workflow.yml | 14 ++++++++++++++ .github/workflows/__swift-custom-build.yml | 14 ++++++++++++++ .github/workflows/__unset-environment.yml | 14 ++++++++++++++ .github/workflows/__upload-ref-sha-input.yml | 14 ++++++++++++++ .github/workflows/__upload-sarif.yml | 14 ++++++++++++++ .github/workflows/__with-checkout-path.yml | 14 ++++++++++++++ pr-checks/checks/all-platform-bundle.yml | 1 + pr-checks/checks/analyze-ref-input.yml | 1 + pr-checks/checks/autobuild-action.yml | 1 + pr-checks/checks/build-mode-manual.yml | 1 + .../export-file-baseline-information.yml | 1 + pr-checks/checks/go-custom-queries.yml | 1 + pr-checks/checks/multi-language-autodetect.yml | 1 + ...packaging-codescanning-config-inputs-js.yml | 1 + .../checks/packaging-config-inputs-js.yml | 1 + pr-checks/checks/packaging-config-js.yml | 1 + pr-checks/checks/packaging-inputs-js.yml | 1 + pr-checks/checks/remote-config.yml | 1 + pr-checks/checks/split-workflow.yml | 1 + pr-checks/checks/swift-custom-build.yml | 1 + pr-checks/checks/unset-environment.yml | 1 + pr-checks/checks/upload-ref-sha-input.yml | 1 + pr-checks/checks/upload-sarif.yml | 1 + pr-checks/checks/with-checkout-path.yml | 1 + 37 files changed, 278 insertions(+), 2 deletions(-) diff --git a/.github/workflows/__all-platform-bundle.yml b/.github/workflows/__all-platform-bundle.yml index 89138c523..e2b5e69fc 100644 --- a/.github/workflows/__all-platform-bundle.yml +++ b/.github/workflows/__all-platform-bundle.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -74,6 +84,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - id: init uses: ./../action/init with: diff --git a/.github/workflows/__analyze-ref-input.yml b/.github/workflows/__analyze-ref-input.yml index 9cf85b0c5..9efe4a8c3 100644 --- a/.github/workflows/__analyze-ref-input.yml +++ b/.github/workflows/__analyze-ref-input.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -85,6 +95,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: tools: ${{ steps.prepare-test.outputs.tools-url }} diff --git a/.github/workflows/__autobuild-action.yml b/.github/workflows/__autobuild-action.yml index c31576339..0e617afe1 100644 --- a/.github/workflows/__autobuild-action.yml +++ b/.github/workflows/__autobuild-action.yml @@ -21,9 +21,19 @@ on: schedule: - cron: '0 5 * * *' workflow_dispatch: - inputs: {} + inputs: + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: - inputs: {} + inputs: + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -59,6 +69,10 @@ jobs: version: ${{ matrix.version }} use-all-platform-bundle: 'false' setup-kotlin: 'true' + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: languages: csharp diff --git a/.github/workflows/__build-mode-manual.yml b/.github/workflows/__build-mode-manual.yml index e0dc25f88..4be0c42d1 100644 --- a/.github/workflows/__build-mode-manual.yml +++ b/.github/workflows/__build-mode-manual.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -70,6 +80,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init id: init with: diff --git a/.github/workflows/__export-file-baseline-information.yml b/.github/workflows/__export-file-baseline-information.yml index 66274f26b..980535c84 100644 --- a/.github/workflows/__export-file-baseline-information.yml +++ b/.github/workflows/__export-file-baseline-information.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -74,6 +84,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init id: init with: diff --git a/.github/workflows/__go-custom-queries.yml b/.github/workflows/__go-custom-queries.yml index 1b5b7b915..fe35b5b4d 100644 --- a/.github/workflows/__go-custom-queries.yml +++ b/.github/workflows/__go-custom-queries.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -72,6 +82,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: languages: go diff --git a/.github/workflows/__go.yml b/.github/workflows/__go.yml index e8bf8837e..fb27da710 100644 --- a/.github/workflows/__go.yml +++ b/.github/workflows/__go.yml @@ -18,6 +18,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x jobs: go-custom-queries: name: 'Go: Custom queries' @@ -27,6 +32,7 @@ jobs: uses: ./.github/workflows/__go-custom-queries.yml with: go-version: ${{ inputs.go-version }} + dotnet-version: ${{ inputs.dotnet-version }} go-indirect-tracing-workaround-diagnostic: name: 'Go: diagnostic when Go is changed after init step' permissions: diff --git a/.github/workflows/__multi-language-autodetect.yml b/.github/workflows/__multi-language-autodetect.yml index e9272f893..3704cdbf5 100644 --- a/.github/workflows/__multi-language-autodetect.yml +++ b/.github/workflows/__multi-language-autodetect.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -119,6 +129,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - name: Use Xcode 16 if: runner.os == 'macOS' && matrix.version != 'nightly-latest' run: sudo xcode-select -s "/Applications/Xcode_16.app" diff --git a/.github/workflows/__packaging-codescanning-config-inputs-js.yml b/.github/workflows/__packaging-codescanning-config-inputs-js.yml index a5038cefc..53f280ab9 100644 --- a/.github/workflows/__packaging-codescanning-config-inputs-js.yml +++ b/.github/workflows/__packaging-codescanning-config-inputs-js.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -96,6 +106,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: config-file: .github/codeql/codeql-config-packaging3.yml diff --git a/.github/workflows/__packaging-config-inputs-js.yml b/.github/workflows/__packaging-config-inputs-js.yml index 9c8483c74..2b483b41a 100644 --- a/.github/workflows/__packaging-config-inputs-js.yml +++ b/.github/workflows/__packaging-config-inputs-js.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -81,6 +91,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: config-file: .github/codeql/codeql-config-packaging3.yml diff --git a/.github/workflows/__packaging-config-js.yml b/.github/workflows/__packaging-config-js.yml index b7608793a..d45ca3b36 100644 --- a/.github/workflows/__packaging-config-js.yml +++ b/.github/workflows/__packaging-config-js.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -81,6 +91,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: config-file: .github/codeql/codeql-config-packaging.yml diff --git a/.github/workflows/__packaging-inputs-js.yml b/.github/workflows/__packaging-inputs-js.yml index 0a814a050..41ca571b8 100644 --- a/.github/workflows/__packaging-inputs-js.yml +++ b/.github/workflows/__packaging-inputs-js.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -81,6 +91,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: config-file: .github/codeql/codeql-config-packaging2.yml diff --git a/.github/workflows/__remote-config.yml b/.github/workflows/__remote-config.yml index b59da417e..20a308e74 100644 --- a/.github/workflows/__remote-config.yml +++ b/.github/workflows/__remote-config.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -87,6 +97,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: tools: ${{ steps.prepare-test.outputs.tools-url }} diff --git a/.github/workflows/__split-workflow.yml b/.github/workflows/__split-workflow.yml index e916b36cc..3ffb09928 100644 --- a/.github/workflows/__split-workflow.yml +++ b/.github/workflows/__split-workflow.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -80,6 +90,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: config-file: .github/codeql/codeql-config-packaging3.yml diff --git a/.github/workflows/__swift-custom-build.yml b/.github/workflows/__swift-custom-build.yml index 32ce33a7f..a1c5a556f 100644 --- a/.github/workflows/__swift-custom-build.yml +++ b/.github/workflows/__swift-custom-build.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -74,6 +84,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - name: Use Xcode 16 if: runner.os == 'macOS' && matrix.version != 'nightly-latest' run: sudo xcode-select -s "/Applications/Xcode_16.app" diff --git a/.github/workflows/__unset-environment.yml b/.github/workflows/__unset-environment.yml index ced3b33fc..c1a62b110 100644 --- a/.github/workflows/__unset-environment.yml +++ b/.github/workflows/__unset-environment.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -87,6 +97,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init id: init with: diff --git a/.github/workflows/__upload-ref-sha-input.yml b/.github/workflows/__upload-ref-sha-input.yml index 8f1fd4917..1c2c5975d 100644 --- a/.github/workflows/__upload-ref-sha-input.yml +++ b/.github/workflows/__upload-ref-sha-input.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -85,6 +95,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: tools: ${{ steps.prepare-test.outputs.tools-url }} diff --git a/.github/workflows/__upload-sarif.yml b/.github/workflows/__upload-sarif.yml index 77cd057f9..361c8228f 100644 --- a/.github/workflows/__upload-sarif.yml +++ b/.github/workflows/__upload-sarif.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -92,6 +102,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: tools: ${{ steps.prepare-test.outputs.tools-url }} diff --git a/.github/workflows/__with-checkout-path.yml b/.github/workflows/__with-checkout-path.yml index 1b4002ffb..5aa2b631c 100644 --- a/.github/workflows/__with-checkout-path.yml +++ b/.github/workflows/__with-checkout-path.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -85,6 +95,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - name: Delete original checkout run: | # delete the original checkout so we don't accidentally use it. diff --git a/pr-checks/checks/all-platform-bundle.yml b/pr-checks/checks/all-platform-bundle.yml index 3396be22a..994c91eb9 100644 --- a/pr-checks/checks/all-platform-bundle.yml +++ b/pr-checks/checks/all-platform-bundle.yml @@ -4,6 +4,7 @@ operatingSystems: ["ubuntu", "macos", "windows"] versions: ["nightly-latest"] useAllPlatformBundle: "true" installGo: true +installDotNet: true steps: - id: init uses: ./../action/init diff --git a/pr-checks/checks/analyze-ref-input.yml b/pr-checks/checks/analyze-ref-input.yml index 1dbf9e0c8..e9d2cd176 100644 --- a/pr-checks/checks/analyze-ref-input.yml +++ b/pr-checks/checks/analyze-ref-input.yml @@ -3,6 +3,7 @@ description: "Checks that specifying 'ref' and 'sha' as inputs works" versions: ["default"] installGo: true installPython: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/autobuild-action.yml b/pr-checks/checks/autobuild-action.yml index 91ae7834c..b91489cc8 100644 --- a/pr-checks/checks/autobuild-action.yml +++ b/pr-checks/checks/autobuild-action.yml @@ -2,6 +2,7 @@ name: "autobuild-action" description: "Tests that the C# autobuild action works" operatingSystems: ["ubuntu", "macos", "windows"] versions: ["linked"] +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/build-mode-manual.yml b/pr-checks/checks/build-mode-manual.yml index f1815b7ff..a8048230c 100644 --- a/pr-checks/checks/build-mode-manual.yml +++ b/pr-checks/checks/build-mode-manual.yml @@ -2,6 +2,7 @@ name: "Build mode manual" description: "An end-to-end integration test of a Java repository built using 'build-mode: manual'" versions: ["nightly-latest"] installGo: true +installDotNet: true steps: - uses: ./../action/init id: init diff --git a/pr-checks/checks/export-file-baseline-information.yml b/pr-checks/checks/export-file-baseline-information.yml index 1a316d6ec..5ff186e9f 100644 --- a/pr-checks/checks/export-file-baseline-information.yml +++ b/pr-checks/checks/export-file-baseline-information.yml @@ -3,6 +3,7 @@ description: "Tests that file baseline information is exported when the feature operatingSystems: ["ubuntu", "macos", "windows"] versions: ["nightly-latest"] installGo: true +installDotNet: true env: CODEQL_ACTION_SUBLANGUAGE_FILE_COVERAGE: true steps: diff --git a/pr-checks/checks/go-custom-queries.yml b/pr-checks/checks/go-custom-queries.yml index ca00fd81a..867d8dc6c 100644 --- a/pr-checks/checks/go-custom-queries.yml +++ b/pr-checks/checks/go-custom-queries.yml @@ -7,6 +7,7 @@ versions: - linked - nightly-latest installGo: true +installDotNet: true env: DOTNET_GENERATE_ASPNET_CERTIFICATE: "false" steps: diff --git a/pr-checks/checks/multi-language-autodetect.yml b/pr-checks/checks/multi-language-autodetect.yml index 84c779dd9..4892bcc31 100644 --- a/pr-checks/checks/multi-language-autodetect.yml +++ b/pr-checks/checks/multi-language-autodetect.yml @@ -5,6 +5,7 @@ env: CODEQL_ACTION_RESOLVE_SUPPORTED_LANGUAGES_USING_CLI: true installGo: true installPython: true +installDotNet: true steps: - name: Use Xcode 16 if: runner.os == 'macOS' && matrix.version != 'nightly-latest' diff --git a/pr-checks/checks/packaging-codescanning-config-inputs-js.yml b/pr-checks/checks/packaging-codescanning-config-inputs-js.yml index 1b72b06b3..6fd0f7c8a 100644 --- a/pr-checks/checks/packaging-codescanning-config-inputs-js.yml +++ b/pr-checks/checks/packaging-codescanning-config-inputs-js.yml @@ -4,6 +4,7 @@ versions: ["linked", "default", "nightly-latest"] # This feature is not compatib installGo: true installNode: true installPython: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/packaging-config-inputs-js.yml b/pr-checks/checks/packaging-config-inputs-js.yml index 41275fd15..8df42f944 100644 --- a/pr-checks/checks/packaging-config-inputs-js.yml +++ b/pr-checks/checks/packaging-config-inputs-js.yml @@ -3,6 +3,7 @@ description: "Checks that specifying packages using a combination of a config fi versions: ["linked", "default", "nightly-latest"] # This feature is not compatible with old CLIs installGo: true installNode: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/packaging-config-js.yml b/pr-checks/checks/packaging-config-js.yml index 906a3a7d9..9fa41061c 100644 --- a/pr-checks/checks/packaging-config-js.yml +++ b/pr-checks/checks/packaging-config-js.yml @@ -3,6 +3,7 @@ description: "Checks that specifying packages using only a config file works" versions: ["linked", "default", "nightly-latest"] # This feature is not compatible with old CLIs installGo: true installNode: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/packaging-inputs-js.yml b/pr-checks/checks/packaging-inputs-js.yml index 9d9fbe71f..bb70de7e6 100644 --- a/pr-checks/checks/packaging-inputs-js.yml +++ b/pr-checks/checks/packaging-inputs-js.yml @@ -3,6 +3,7 @@ description: "Checks that specifying packages using the input to the Action work versions: ["linked", "default", "nightly-latest"] # This feature is not compatible with old CLIs installGo: true installNode: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/remote-config.yml b/pr-checks/checks/remote-config.yml index c0a9cf0bb..24249156e 100644 --- a/pr-checks/checks/remote-config.yml +++ b/pr-checks/checks/remote-config.yml @@ -7,6 +7,7 @@ versions: - nightly-latest installGo: true installPython: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/split-workflow.yml b/pr-checks/checks/split-workflow.yml index fdcf1d530..23f82a7a5 100644 --- a/pr-checks/checks/split-workflow.yml +++ b/pr-checks/checks/split-workflow.yml @@ -3,6 +3,7 @@ description: "Tests a split-up workflow in which we first build a database and l operatingSystems: ["ubuntu", "macos"] versions: ["linked", "default", "nightly-latest"] # This feature is not compatible with old CLIs installGo: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/swift-custom-build.yml b/pr-checks/checks/swift-custom-build.yml index 2ad44ff3b..1c8f1bf3a 100644 --- a/pr-checks/checks/swift-custom-build.yml +++ b/pr-checks/checks/swift-custom-build.yml @@ -3,6 +3,7 @@ description: "Tests creation of a Swift database using custom build" versions: ["linked", "default", "nightly-latest"] operatingSystems: ["macos"] installGo: true +installDotNet: true env: DOTNET_GENERATE_ASPNET_CERTIFICATE: "false" steps: diff --git a/pr-checks/checks/unset-environment.yml b/pr-checks/checks/unset-environment.yml index 614c2006e..4cc728600 100644 --- a/pr-checks/checks/unset-environment.yml +++ b/pr-checks/checks/unset-environment.yml @@ -7,6 +7,7 @@ versions: - nightly-latest installGo: true installPython: true +installDotNet: true steps: - uses: ./../action/init id: init diff --git a/pr-checks/checks/upload-ref-sha-input.yml b/pr-checks/checks/upload-ref-sha-input.yml index fe55b5457..0c8059a51 100644 --- a/pr-checks/checks/upload-ref-sha-input.yml +++ b/pr-checks/checks/upload-ref-sha-input.yml @@ -3,6 +3,7 @@ description: "Checks that specifying 'ref' and 'sha' as inputs works" versions: ["default"] installGo: true installPython: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/upload-sarif.yml b/pr-checks/checks/upload-sarif.yml index 0028e67ed..cfe66a3f8 100644 --- a/pr-checks/checks/upload-sarif.yml +++ b/pr-checks/checks/upload-sarif.yml @@ -4,6 +4,7 @@ versions: ["default"] analysisKinds: ["code-scanning", "code-quality", "code-scanning,code-quality"] installGo: true installPython: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/with-checkout-path.yml b/pr-checks/checks/with-checkout-path.yml index b18d191cc..5cdd02c0d 100644 --- a/pr-checks/checks/with-checkout-path.yml +++ b/pr-checks/checks/with-checkout-path.yml @@ -3,6 +3,7 @@ description: "Checks that a custom `checkout_path` will find the proper commit_o versions: ["linked"] installGo: true installPython: true +installDotNet: true steps: # This ensures we don't accidentally use the original checkout for any part of the test. - name: Delete original checkout From 3fac49c14012959fab197865a1813bffe8a04dd1 Mon Sep 17 00:00:00 2001 From: "Michael B. Gale" Date: Thu, 13 Nov 2025 19:53:24 +0000 Subject: [PATCH 07/51] Update remaining workflows --- .github/workflows/debug-artifacts-failure-safe.yml | 4 ++++ .github/workflows/debug-artifacts-safe.yml | 4 ++++ .github/workflows/test-codeql-bundle-all.yml | 4 ++++ 3 files changed, 12 insertions(+) diff --git a/.github/workflows/debug-artifacts-failure-safe.yml b/.github/workflows/debug-artifacts-failure-safe.yml index 1a09b3d9e..768f88f96 100644 --- a/.github/workflows/debug-artifacts-failure-safe.yml +++ b/.github/workflows/debug-artifacts-failure-safe.yml @@ -54,6 +54,10 @@ jobs: - uses: actions/setup-go@v6 with: go-version: ^1.13.1 + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: '9.x' - uses: ./../action/init with: tools: ${{ steps.prepare-test.outputs.tools-url }} diff --git a/.github/workflows/debug-artifacts-safe.yml b/.github/workflows/debug-artifacts-safe.yml index ea513521f..e33d70cc3 100644 --- a/.github/workflows/debug-artifacts-safe.yml +++ b/.github/workflows/debug-artifacts-safe.yml @@ -50,6 +50,10 @@ jobs: - uses: actions/setup-go@v6 with: go-version: ^1.13.1 + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: '9.x' - uses: ./../action/init id: init with: diff --git a/.github/workflows/test-codeql-bundle-all.yml b/.github/workflows/test-codeql-bundle-all.yml index 4b7fdca81..6465d6a1d 100644 --- a/.github/workflows/test-codeql-bundle-all.yml +++ b/.github/workflows/test-codeql-bundle-all.yml @@ -43,6 +43,10 @@ jobs: with: version: ${{ matrix.version }} use-all-platform-bundle: true + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: '9.x' - id: init uses: ./../action/init with: From 14d898ef09787f3258d7542ed62ad3da38295b68 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Thu, 13 Nov 2025 21:18:01 +0000 Subject: [PATCH 08/51] Update changelog for v4.31.3 --- CHANGELOG.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 923cc8264..ff1d40d36 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -2,7 +2,7 @@ See the [releases page](https://github.com/github/codeql-action/releases) for the relevant changes to the CodeQL CLI and language packs. -## [UNRELEASED] +## 4.31.3 - 13 Nov 2025 - CodeQL Action v3 will be deprecated in December 2026. The Action now logs a warning for customers who are running v3 but could be running v4. For more information, see [Upcoming deprecation of CodeQL Action v3](https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/). - Update default CodeQL bundle version to 2.23.5. [#3288](https://github.com/github/codeql-action/pull/3288) From 497c7f627a731a7a484b2a1341bd4b5ff658f48b Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Thu, 13 Nov 2025 21:54:56 +0000 Subject: [PATCH 09/51] Update changelog and version after v4.31.3 --- CHANGELOG.md | 4 ++++ package-lock.json | 4 ++-- package.json | 2 +- 3 files changed, 7 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index ff1d40d36..4ccc60273 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -2,6 +2,10 @@ See the [releases page](https://github.com/github/codeql-action/releases) for the relevant changes to the CodeQL CLI and language packs. +## [UNRELEASED] + +No user facing changes. + ## 4.31.3 - 13 Nov 2025 - CodeQL Action v3 will be deprecated in December 2026. The Action now logs a warning for customers who are running v3 but could be running v4. For more information, see [Upcoming deprecation of CodeQL Action v3](https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/). diff --git a/package-lock.json b/package-lock.json index 78a1b05a6..28e08ed18 100644 --- a/package-lock.json +++ b/package-lock.json @@ -1,12 +1,12 @@ { "name": "codeql", - "version": "4.31.3", + "version": "4.31.4", "lockfileVersion": 3, "requires": true, "packages": { "": { "name": "codeql", - "version": "4.31.3", + "version": "4.31.4", "license": "MIT", "dependencies": { "@actions/artifact": "^4.0.0", diff --git a/package.json b/package.json index dcad231e2..c7162d9c4 100644 --- a/package.json +++ b/package.json @@ -1,6 +1,6 @@ { "name": "codeql", - "version": "4.31.3", + "version": "4.31.4", "private": true, "description": "CodeQL action", "scripts": { From 246edb9b1d0b05118e62df6777f757001e3614e1 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Thu, 13 Nov 2025 21:59:57 +0000 Subject: [PATCH 10/51] Rebuild --- lib/analyze-action-post.js | 2 +- lib/analyze-action.js | 2 +- lib/autobuild-action.js | 2 +- lib/init-action-post.js | 2 +- lib/init-action.js | 2 +- lib/resolve-environment-action.js | 2 +- lib/setup-codeql-action.js | 2 +- lib/start-proxy-action-post.js | 2 +- lib/start-proxy-action.js | 2 +- lib/upload-lib.js | 2 +- lib/upload-sarif-action-post.js | 2 +- lib/upload-sarif-action.js | 2 +- 12 files changed, 12 insertions(+), 12 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index b30ab9f09..5145c8bc2 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.4", private: true, description: "CodeQL action", scripts: { diff --git a/lib/analyze-action.js b/lib/analyze-action.js index deb4af906..4ee0d87e5 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.4", private: true, description: "CodeQL action", scripts: { diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index fdeb4d70d..8ff3dd6cd 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.4", private: true, description: "CodeQL action", scripts: { diff --git a/lib/init-action-post.js b/lib/init-action-post.js index ef36db529..af32a25c1 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.4", private: true, description: "CodeQL action", scripts: { diff --git a/lib/init-action.js b/lib/init-action.js index 41cdafba1..1bd15d9b7 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.4", private: true, description: "CodeQL action", scripts: { diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index 7918ab61f..ff90d9c4e 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.4", private: true, description: "CodeQL action", scripts: { diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index e119b9477..9cbb52ad3 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.4", private: true, description: "CodeQL action", scripts: { diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index 2386e7c27..2f1d4544a 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.4", private: true, description: "CodeQL action", scripts: { diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index 2af00a59a..d51ba32df 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -47285,7 +47285,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.4", private: true, description: "CodeQL action", scripts: { diff --git a/lib/upload-lib.js b/lib/upload-lib.js index dc7d0d162..fb897b861 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -28924,7 +28924,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.4", private: true, description: "CodeQL action", scripts: { diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index 1d2a3a44b..5a9cf35a2 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.4", private: true, description: "CodeQL action", scripts: { diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 46c84e88c..d0b2dd324 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.4", private: true, description: "CodeQL action", scripts: { From 11889c27fd4014f9106e8f91eba5989f7b84f222 Mon Sep 17 00:00:00 2001 From: "Michael B. Gale" Date: Fri, 14 Nov 2025 11:36:59 +0000 Subject: [PATCH 11/51] Return keys of restored caches from `downloadDependencyCaches` --- lib/init-action.js | 24 +++++++++---- src/dependency-caching.test.ts | 62 ++++++++++++++++++++++++++-------- src/dependency-caching.ts | 30 +++++++++++++--- src/init-action.ts | 9 ++--- 4 files changed, 95 insertions(+), 30 deletions(-) diff --git a/lib/init-action.js b/lib/init-action.js index 6c92b83a7..aaa4836d7 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -87317,6 +87317,7 @@ async function checkHashPatterns(codeql, features, language, cacheConfig, checkT } async function downloadDependencyCaches(codeql, features, languages, logger) { const status = []; + const restoredKeys = []; for (const language of languages) { const cacheConfig = defaultCacheConfigs[language]; if (cacheConfig === void 0) { @@ -87355,14 +87356,22 @@ async function downloadDependencyCaches(codeql, features, languages, logger) { const download_duration_ms = Math.round(performance.now() - start); if (hitKey !== void 0) { logger.info(`Cache hit on key ${hitKey} for ${language}.`); - const hit_kind = hitKey === primaryKey ? "exact" /* Exact */ : "partial" /* Partial */; - status.push({ language, hit_kind, download_duration_ms }); + let hit_kind = "partial" /* Partial */; + if (hitKey === primaryKey) { + hit_kind = "exact" /* Exact */; + } + status.push({ + language, + hit_kind, + download_duration_ms + }); + restoredKeys.push(hitKey); } else { status.push({ language, hit_kind: "miss" /* Miss */ }); logger.info(`No suitable cache found for ${language}.`); } } - return status; + return { statusReport: status, restoredKeys }; } async function cacheKey2(codeql, features, language, patterns) { const hash = await glob.hashFiles(patterns.join("\n")); @@ -89994,7 +90003,7 @@ async function run() { return; } let overlayBaseDatabaseStats; - let dependencyCachingResults; + let dependencyCachingStatus; try { if (config.overlayDatabaseMode === "overlay" /* Overlay */ && config.useOverlayDatabaseCaching) { overlayBaseDatabaseStats = await downloadOverlayBaseDatabaseFromCache( @@ -90135,12 +90144,13 @@ exec ${goBinaryPath} "$@"` } } if (shouldRestoreCache(config.dependencyCachingEnabled)) { - dependencyCachingResults = await downloadDependencyCaches( + const dependencyCachingResult = await downloadDependencyCaches( codeql, features, config.languages, logger ); + dependencyCachingStatus = dependencyCachingResult.statusReport; } if (await codeQlVersionAtLeast(codeql, "2.17.1")) { } else { @@ -90241,7 +90251,7 @@ exec ${goBinaryPath} "$@"` toolsSource, toolsVersion, overlayBaseDatabaseStats, - dependencyCachingResults, + dependencyCachingStatus, logger, error4 ); @@ -90259,7 +90269,7 @@ exec ${goBinaryPath} "$@"` toolsSource, toolsVersion, overlayBaseDatabaseStats, - dependencyCachingResults, + dependencyCachingStatus, logger ); } diff --git a/src/dependency-caching.test.ts b/src/dependency-caching.test.ts index eefb8504c..3c8f0f276 100644 --- a/src/dependency-caching.test.ts +++ b/src/dependency-caching.test.ts @@ -237,15 +237,17 @@ test("downloadDependencyCaches - does not restore caches with feature keys if no .resolves(CSHARP_BASE_PATTERNS); makePatternCheckStub.withArgs(CSHARP_EXTRA_PATTERNS).resolves(undefined); - const results = await downloadDependencyCaches( + const result = await downloadDependencyCaches( codeql, createFeatures([]), [KnownLanguage.csharp], logger, ); - t.is(results.length, 1); - t.is(results[0].language, KnownLanguage.csharp); - t.is(results[0].hit_kind, CacheHitKind.Miss); + const statusReport = result.statusReport; + t.is(statusReport.length, 1); + t.is(statusReport[0].language, KnownLanguage.csharp); + t.is(statusReport[0].hit_kind, CacheHitKind.Miss); + t.deepEqual(result.restoredKeys, []); t.assert(restoreCacheStub.calledOnce); }); @@ -257,7 +259,8 @@ test("downloadDependencyCaches - restores caches with feature keys if features a const logger = getRecordingLogger(messages); const features = createFeatures([Feature.CsharpNewCacheKey]); - sinon.stub(glob, "hashFiles").resolves("abcdef"); + const mockHash = "abcdef"; + sinon.stub(glob, "hashFiles").resolves(mockHash); const keyWithFeature = await cacheKey( codeql, @@ -277,15 +280,28 @@ test("downloadDependencyCaches - restores caches with feature keys if features a .resolves(CSHARP_BASE_PATTERNS); makePatternCheckStub.withArgs(CSHARP_EXTRA_PATTERNS).resolves(undefined); - const results = await downloadDependencyCaches( + const result = await downloadDependencyCaches( codeql, features, [KnownLanguage.csharp], logger, ); - t.is(results.length, 1); - t.is(results[0].language, KnownLanguage.csharp); - t.is(results[0].hit_kind, CacheHitKind.Exact); + + // Check that the status report for telemetry indicates that one cache was restored with an exact match. + const statusReport = result.statusReport; + t.is(statusReport.length, 1); + t.is(statusReport[0].language, KnownLanguage.csharp); + t.is(statusReport[0].hit_kind, CacheHitKind.Exact); + + // Check that the restored key has been returned. + const restoredKeys = result.restoredKeys; + t.is(restoredKeys.length, 1); + t.assert( + restoredKeys[0].endsWith(mockHash), + "Expected restored key to end with hash returned by `hashFiles`", + ); + + // `restoreCache` should have been called exactly once. t.assert(restoreCacheStub.calledOnce); }); @@ -297,8 +313,14 @@ test("downloadDependencyCaches - restores caches with feature keys if features a const logger = getRecordingLogger(messages); const features = createFeatures([Feature.CsharpNewCacheKey]); + // We expect two calls to `hashFiles`: the first by the call to `cacheKey` below, + // and the second by `downloadDependencyCaches`. We use the result of the first + // call as part of the cache key that identifies a mock, existing cache. The result + // of the second call is for the primary restore key, which we don't want to match + // the first key so that we can test the restore keys logic. + const restoredHash = "abcdef"; const hashFilesStub = sinon.stub(glob, "hashFiles"); - hashFilesStub.onFirstCall().resolves("abcdef"); + hashFilesStub.onFirstCall().resolves(restoredHash); hashFilesStub.onSecondCall().resolves("123456"); const keyWithFeature = await cacheKey( @@ -319,15 +341,27 @@ test("downloadDependencyCaches - restores caches with feature keys if features a .resolves(CSHARP_BASE_PATTERNS); makePatternCheckStub.withArgs(CSHARP_EXTRA_PATTERNS).resolves(undefined); - const results = await downloadDependencyCaches( + const result = await downloadDependencyCaches( codeql, features, [KnownLanguage.csharp], logger, ); - t.is(results.length, 1); - t.is(results[0].language, KnownLanguage.csharp); - t.is(results[0].hit_kind, CacheHitKind.Partial); + + // Check that the status report for telemetry indicates that one cache was restored with a partial match. + const statusReport = result.statusReport; + t.is(statusReport.length, 1); + t.is(statusReport[0].language, KnownLanguage.csharp); + t.is(statusReport[0].hit_kind, CacheHitKind.Partial); + + // Check that the restored key has been returned. + const restoredKeys = result.restoredKeys; + t.is(restoredKeys.length, 1); + t.assert( + restoredKeys[0].endsWith(restoredHash), + "Expected restored key to end with hash returned by `hashFiles`", + ); + t.assert(restoreCacheStub.calledOnce); }); diff --git a/src/dependency-caching.ts b/src/dependency-caching.ts index 220f1d5ba..22c5b40b3 100644 --- a/src/dependency-caching.ts +++ b/src/dependency-caching.ts @@ -193,6 +193,14 @@ export interface DependencyCacheRestoreStatus { /** An array of `DependencyCacheRestoreStatus` objects for each analysed language with a caching configuration. */ export type DependencyCacheRestoreStatusReport = DependencyCacheRestoreStatus[]; +/** Represents the results of `downloadDependencyCaches`. */ +export interface DownloadDependencyCachesResult { + /** The status report for telemetry */ + statusReport: DependencyCacheRestoreStatusReport; + /** An array of cache keys that we have restored and therefore know to exist. */ + restoredKeys: string[]; +} + /** * A wrapper around `cacheConfig.getHashPatterns` which logs when there are no files to calculate * a hash for the cache key from. @@ -239,8 +247,9 @@ export async function downloadDependencyCaches( features: FeatureEnablement, languages: Language[], logger: Logger, -): Promise { +): Promise { const status: DependencyCacheRestoreStatusReport = []; + const restoredKeys: string[] = []; for (const language of languages) { const cacheConfig = defaultCacheConfigs[language]; @@ -288,16 +297,27 @@ export async function downloadDependencyCaches( if (hitKey !== undefined) { logger.info(`Cache hit on key ${hitKey} for ${language}.`); - const hit_kind = - hitKey === primaryKey ? CacheHitKind.Exact : CacheHitKind.Partial; - status.push({ language, hit_kind, download_duration_ms }); + + // We have a partial cache hit, unless the key of the restored cache matches the + // primary restore key. + let hit_kind = CacheHitKind.Partial; + if (hitKey === primaryKey) { + hit_kind = CacheHitKind.Exact; + } + + status.push({ + language, + hit_kind, + download_duration_ms, + }); + restoredKeys.push(hitKey); } else { status.push({ language, hit_kind: CacheHitKind.Miss }); logger.info(`No suitable cache found for ${language}.`); } } - return status; + return { statusReport: status, restoredKeys }; } /** Enumerates possible outcomes for storing caches. */ diff --git a/src/init-action.ts b/src/init-action.ts index 3512520c2..9f1c33e68 100644 --- a/src/init-action.ts +++ b/src/init-action.ts @@ -371,7 +371,7 @@ async function run() { } let overlayBaseDatabaseStats: OverlayBaseDatabaseDownloadStats | undefined; - let dependencyCachingResults: DependencyCacheRestoreStatusReport | undefined; + let dependencyCachingStatus: DependencyCacheRestoreStatusReport | undefined; try { if ( config.overlayDatabaseMode === OverlayDatabaseMode.Overlay && @@ -579,12 +579,13 @@ async function run() { // Restore dependency cache(s), if they exist. if (shouldRestoreCache(config.dependencyCachingEnabled)) { - dependencyCachingResults = await downloadDependencyCaches( + const dependencyCachingResult = await downloadDependencyCaches( codeql, features, config.languages, logger, ); + dependencyCachingStatus = dependencyCachingResult.statusReport; } // Suppress warnings about disabled Python library extraction. @@ -732,7 +733,7 @@ async function run() { toolsSource, toolsVersion, overlayBaseDatabaseStats, - dependencyCachingResults, + dependencyCachingStatus, logger, error, ); @@ -755,7 +756,7 @@ async function run() { toolsSource, toolsVersion, overlayBaseDatabaseStats, - dependencyCachingResults, + dependencyCachingStatus, logger, ); } From 594c0cc369ddc18703f37fc856dc03cb56c60736 Mon Sep 17 00:00:00 2001 From: "Michael B. Gale" Date: Fri, 14 Nov 2025 11:49:24 +0000 Subject: [PATCH 12/51] Store restored keys in action state --- lib/init-action.js | 2 ++ src/config-utils.test.ts | 27 ++++----------------------- src/config-utils.ts | 4 ++++ src/init-action.ts | 2 ++ src/testing-utils.ts | 1 + 5 files changed, 13 insertions(+), 23 deletions(-) diff --git a/lib/init-action.js b/lib/init-action.js index aaa4836d7..5dad5fc06 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -86824,6 +86824,7 @@ async function initActionState({ trapCaches, trapCacheDownloadTime, dependencyCachingEnabled: getCachingKind(dependencyCachingEnabled), + dependencyCachingRestoredKeys: [], extraQueryExclusions: [], overlayDatabaseMode: "none" /* None */, useOverlayDatabaseCaching: false, @@ -90151,6 +90152,7 @@ exec ${goBinaryPath} "$@"` logger ); dependencyCachingStatus = dependencyCachingResult.statusReport; + config.dependencyCachingRestoredKeys = dependencyCachingResult.restoredKeys; } if (await codeQlVersionAtLeast(codeql, "2.17.1")) { } else { diff --git a/src/config-utils.test.ts b/src/config-utils.test.ts index da58dd8b1..906336d26 100644 --- a/src/config-utils.test.ts +++ b/src/config-utils.test.ts @@ -200,12 +200,9 @@ test("load code quality config", async (t) => { ); // And the config we expect it to result in - const expectedConfig: configUtils.Config = { - version: actionsUtil.getActionVersion(), + const expectedConfig = createTestConfig({ analysisKinds: [AnalysisKind.CodeQuality], languages: [KnownLanguage.actions], - buildMode: undefined, - originalUserInput: {}, // This gets set because we only have `AnalysisKind.CodeQuality` computedConfig: { "disable-default-queries": true, @@ -219,14 +216,7 @@ test("load code quality config", async (t) => { debugMode: false, debugArtifactName: "", debugDatabaseName: "", - trapCaches: {}, - trapCacheDownloadTime: 0, - dependencyCachingEnabled: CachingKind.None, - extraQueryExclusions: [], - overlayDatabaseMode: OverlayDatabaseMode.None, - useOverlayDatabaseCaching: false, - repositoryProperties: {}, - }; + }); t.deepEqual(config, expectedConfig); }); @@ -507,9 +497,7 @@ test("load non-empty input", async (t) => { }; // And the config we expect it to parse to - const expectedConfig: configUtils.Config = { - version: actionsUtil.getActionVersion(), - analysisKinds: [AnalysisKind.CodeScanning], + const expectedConfig = createTestConfig({ languages: [KnownLanguage.javascript], buildMode: BuildMode.None, originalUserInput: userConfig, @@ -521,14 +509,7 @@ test("load non-empty input", async (t) => { debugMode: false, debugArtifactName: "my-artifact", debugDatabaseName: "my-db", - trapCaches: {}, - trapCacheDownloadTime: 0, - dependencyCachingEnabled: CachingKind.None, - extraQueryExclusions: [], - overlayDatabaseMode: OverlayDatabaseMode.None, - useOverlayDatabaseCaching: false, - repositoryProperties: {}, - }; + }); const languagesInput = "javascript"; const configFilePath = createConfigFile(inputFileContents, tempDir); diff --git a/src/config-utils.ts b/src/config-utils.ts index fa88d4b4a..03608fb1a 100644 --- a/src/config-utils.ts +++ b/src/config-utils.ts @@ -148,6 +148,9 @@ export interface Config { /** A value indicating how dependency caching should be used. */ dependencyCachingEnabled: CachingKind; + /** The keys of caches that we restored, if any. */ + dependencyCachingRestoredKeys: string[]; + /** * Extra query exclusions to append to the config. */ @@ -496,6 +499,7 @@ export async function initActionState( trapCaches, trapCacheDownloadTime, dependencyCachingEnabled: getCachingKind(dependencyCachingEnabled), + dependencyCachingRestoredKeys: [], extraQueryExclusions: [], overlayDatabaseMode: OverlayDatabaseMode.None, useOverlayDatabaseCaching: false, diff --git a/src/init-action.ts b/src/init-action.ts index 9f1c33e68..689ded2fc 100644 --- a/src/init-action.ts +++ b/src/init-action.ts @@ -586,6 +586,8 @@ async function run() { logger, ); dependencyCachingStatus = dependencyCachingResult.statusReport; + config.dependencyCachingRestoredKeys = + dependencyCachingResult.restoredKeys; } // Suppress warnings about disabled Python library extraction. diff --git a/src/testing-utils.ts b/src/testing-utils.ts index 12193309b..eaec11e2c 100644 --- a/src/testing-utils.ts +++ b/src/testing-utils.ts @@ -392,6 +392,7 @@ export function createTestConfig(overrides: Partial): Config { trapCaches: {}, trapCacheDownloadTime: 0, dependencyCachingEnabled: CachingKind.None, + dependencyCachingRestoredKeys: [], extraQueryExclusions: [], overlayDatabaseMode: OverlayDatabaseMode.None, useOverlayDatabaseCaching: false, From 51c9af3a3b0efcf59cfdeca89f6bac998d5c0679 Mon Sep 17 00:00:00 2001 From: "Michael B. Gale" Date: Fri, 14 Nov 2025 12:10:38 +0000 Subject: [PATCH 13/51] Don't try to upload cache if we have restored a cache with the same key --- lib/analyze-action.js | 6 +++++- src/dependency-caching.ts | 14 ++++++++++++-- 2 files changed, 17 insertions(+), 3 deletions(-) diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 4ee0d87e5..6453b000f 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -91155,6 +91155,11 @@ async function uploadDependencyCaches(codeql, features, config, logger) { status.push({ language, result: "no-hash" /* NoHash */ }); continue; } + const key = await cacheKey2(codeql, features, language, patterns); + if (config.dependencyCachingRestoredKeys.includes(key)) { + status.push({ language, result: "duplicate" /* Duplicate */ }); + continue; + } const size = await getTotalCacheSize( cacheConfig.getDependencyPaths(), logger, @@ -91167,7 +91172,6 @@ async function uploadDependencyCaches(codeql, features, config, logger) { ); continue; } - const key = await cacheKey2(codeql, features, language, patterns); logger.info( `Uploading cache of size ${size} for ${language} with key ${key}...` ); diff --git a/src/dependency-caching.ts b/src/dependency-caching.ts index 22c5b40b3..9f6aa2afc 100644 --- a/src/dependency-caching.ts +++ b/src/dependency-caching.ts @@ -385,6 +385,18 @@ export async function uploadDependencyCaches( continue; } + // Now that we have verified that there are suitable files, compute the hash for the cache key. + const key = await cacheKey(codeql, features, language, patterns); + + // Check that we haven't previously restored this exact key. If a cache with this key + // already exists in the Actions Cache, performing the next steps is pointless as the cache + // will not get overwritten. We can therefore skip the expensive work of measuring the size + // of the cache contents and attempting to upload it if we know that the cache already exists. + if (config.dependencyCachingRestoredKeys.includes(key)) { + status.push({ language, result: CacheStoreResult.Duplicate }); + continue; + } + // Calculate the size of the files that we would store in the cache. We use this to determine whether the // cache should be saved or not. For example, if there are no files to store, then we skip creating the // cache. In the future, we could also: @@ -410,8 +422,6 @@ export async function uploadDependencyCaches( continue; } - const key = await cacheKey(codeql, features, language, patterns); - logger.info( `Uploading cache of size ${size} for ${language} with key ${key}...`, ); From 1ed85b450113c9629b9679dc18aee3121374c031 Mon Sep 17 00:00:00 2001 From: "Michael B. Gale" Date: Fri, 14 Nov 2025 12:22:23 +0000 Subject: [PATCH 14/51] Add test coverage for `uploadDependencyCaches` --- src/dependency-caching.test.ts | 204 +++++++++++++++++++++++++++++++++ 1 file changed, 204 insertions(+) diff --git a/src/dependency-caching.test.ts b/src/dependency-caching.test.ts index 3c8f0f276..a9b9e6210 100644 --- a/src/dependency-caching.test.ts +++ b/src/dependency-caching.test.ts @@ -7,6 +7,7 @@ import test from "ava"; import * as sinon from "sinon"; import { cacheKeyHashLength } from "./caching-utils"; +import * as cachingUtils from "./caching-utils"; import { createStubCodeQL } from "./codeql"; import { CacheConfig, @@ -20,6 +21,8 @@ import { downloadDependencyCaches, CacheHitKind, cacheKey, + uploadDependencyCaches, + CacheStoreResult, } from "./dependency-caching"; import { Feature } from "./feature-flags"; import { KnownLanguage } from "./languages"; @@ -29,6 +32,7 @@ import { getRecordingLogger, checkExpectedLogMessages, LoggedMessage, + createTestConfig, } from "./testing-utils"; import { withTmpDir } from "./util"; @@ -365,6 +369,206 @@ test("downloadDependencyCaches - restores caches with feature keys if features a t.assert(restoreCacheStub.calledOnce); }); +test("uploadDependencyCaches - skips upload for a language with no cache config", async (t) => { + const codeql = createStubCodeQL({}); + const messages: LoggedMessage[] = []; + const logger = getRecordingLogger(messages); + const features = createFeatures([]); + const config = createTestConfig({ + languages: [KnownLanguage.actions], + }); + + const result = await uploadDependencyCaches(codeql, features, config, logger); + t.is(result.length, 0); + checkExpectedLogMessages(t, messages, [ + "Skipping upload of dependency cache for actions", + ]); +}); + +test("uploadDependencyCaches - skips upload if no files for the hash exist", async (t) => { + const codeql = createStubCodeQL({}); + const messages: LoggedMessage[] = []; + const logger = getRecordingLogger(messages); + const features = createFeatures([]); + const config = createTestConfig({ + languages: [KnownLanguage.go], + }); + + const makePatternCheckStub = sinon.stub(internal, "makePatternCheck"); + makePatternCheckStub.resolves(undefined); + + const result = await uploadDependencyCaches(codeql, features, config, logger); + t.is(result.length, 1); + t.is(result[0].language, KnownLanguage.go); + t.is(result[0].result, CacheStoreResult.NoHash); +}); + +test("uploadDependencyCaches - skips upload if we know the cache already exists", async (t) => { + process.env["RUNNER_OS"] = "Linux"; + + const codeql = createStubCodeQL({}); + const messages: LoggedMessage[] = []; + const logger = getRecordingLogger(messages); + const features = createFeatures([]); + + const mockHash = "abcdef"; + sinon.stub(glob, "hashFiles").resolves(mockHash); + + const makePatternCheckStub = sinon.stub(internal, "makePatternCheck"); + makePatternCheckStub + .withArgs(CSHARP_BASE_PATTERNS) + .resolves(CSHARP_BASE_PATTERNS); + + const primaryCacheKey = await cacheKey( + codeql, + features, + KnownLanguage.csharp, + CSHARP_BASE_PATTERNS, + ); + + const config = createTestConfig({ + languages: [KnownLanguage.csharp], + dependencyCachingRestoredKeys: [primaryCacheKey], + }); + + const result = await uploadDependencyCaches(codeql, features, config, logger); + t.is(result.length, 1); + t.is(result[0].language, KnownLanguage.csharp); + t.is(result[0].result, CacheStoreResult.Duplicate); +}); + +test("uploadDependencyCaches - skips upload if cache size is 0", async (t) => { + process.env["RUNNER_OS"] = "Linux"; + + const codeql = createStubCodeQL({}); + const messages: LoggedMessage[] = []; + const logger = getRecordingLogger(messages); + const features = createFeatures([]); + + const mockHash = "abcdef"; + sinon.stub(glob, "hashFiles").resolves(mockHash); + + const makePatternCheckStub = sinon.stub(internal, "makePatternCheck"); + makePatternCheckStub + .withArgs(CSHARP_BASE_PATTERNS) + .resolves(CSHARP_BASE_PATTERNS); + + sinon.stub(cachingUtils, "getTotalCacheSize").resolves(0); + + const config = createTestConfig({ + languages: [KnownLanguage.csharp], + }); + + const result = await uploadDependencyCaches(codeql, features, config, logger); + t.is(result.length, 1); + t.is(result[0].language, KnownLanguage.csharp); + t.is(result[0].result, CacheStoreResult.Empty); + + checkExpectedLogMessages(t, messages, [ + "Skipping upload of dependency cache", + ]); +}); + +test("uploadDependencyCaches - uploads caches when all requirements are met", async (t) => { + process.env["RUNNER_OS"] = "Linux"; + + const codeql = createStubCodeQL({}); + const messages: LoggedMessage[] = []; + const logger = getRecordingLogger(messages); + const features = createFeatures([]); + + const mockHash = "abcdef"; + sinon.stub(glob, "hashFiles").resolves(mockHash); + + const makePatternCheckStub = sinon.stub(internal, "makePatternCheck"); + makePatternCheckStub + .withArgs(CSHARP_BASE_PATTERNS) + .resolves(CSHARP_BASE_PATTERNS); + + sinon.stub(cachingUtils, "getTotalCacheSize").resolves(1024); + sinon.stub(actionsCache, "saveCache").resolves(); + + const config = createTestConfig({ + languages: [KnownLanguage.csharp], + }); + + const result = await uploadDependencyCaches(codeql, features, config, logger); + t.is(result.length, 1); + t.is(result[0].language, KnownLanguage.csharp); + t.is(result[0].result, CacheStoreResult.Stored); + t.is(result[0].upload_size_bytes, 1024); + + checkExpectedLogMessages(t, messages, ["Uploading cache of size"]); +}); + +test("uploadDependencyCaches - catches `ReserveCacheError` exceptions", async (t) => { + process.env["RUNNER_OS"] = "Linux"; + + const codeql = createStubCodeQL({}); + const messages: LoggedMessage[] = []; + const logger = getRecordingLogger(messages); + const features = createFeatures([]); + + const mockHash = "abcdef"; + sinon.stub(glob, "hashFiles").resolves(mockHash); + + const makePatternCheckStub = sinon.stub(internal, "makePatternCheck"); + makePatternCheckStub + .withArgs(CSHARP_BASE_PATTERNS) + .resolves(CSHARP_BASE_PATTERNS); + + sinon.stub(cachingUtils, "getTotalCacheSize").resolves(1024); + sinon + .stub(actionsCache, "saveCache") + .throws(new actionsCache.ReserveCacheError("Already in use")); + + const config = createTestConfig({ + languages: [KnownLanguage.csharp], + }); + + await t.notThrowsAsync(async () => { + const result = await uploadDependencyCaches( + codeql, + features, + config, + logger, + ); + t.is(result.length, 1); + t.is(result[0].language, KnownLanguage.csharp); + t.is(result[0].result, CacheStoreResult.Duplicate); + + checkExpectedLogMessages(t, messages, ["Not uploading cache for"]); + }); +}); + +test("uploadDependencyCaches - throws other exceptions", async (t) => { + process.env["RUNNER_OS"] = "Linux"; + + const codeql = createStubCodeQL({}); + const messages: LoggedMessage[] = []; + const logger = getRecordingLogger(messages); + const features = createFeatures([]); + + const mockHash = "abcdef"; + sinon.stub(glob, "hashFiles").resolves(mockHash); + + const makePatternCheckStub = sinon.stub(internal, "makePatternCheck"); + makePatternCheckStub + .withArgs(CSHARP_BASE_PATTERNS) + .resolves(CSHARP_BASE_PATTERNS); + + sinon.stub(cachingUtils, "getTotalCacheSize").resolves(1024); + sinon.stub(actionsCache, "saveCache").throws(); + + const config = createTestConfig({ + languages: [KnownLanguage.csharp], + }); + + await t.throwsAsync(async () => { + await uploadDependencyCaches(codeql, features, config, logger); + }); +}); + test("getFeaturePrefix - returns empty string if no features are enabled", async (t) => { const codeql = createStubCodeQL({}); const features = createFeatures([]); From b9620e12499afd2c9b1e79411670cfbfdf83a06e Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Sat, 15 Nov 2025 10:55:57 +0000 Subject: [PATCH 15/51] Bump js-yaml from 4.1.0 to 4.1.1 Bumps [js-yaml](https://github.com/nodeca/js-yaml) from 4.1.0 to 4.1.1. - [Changelog](https://github.com/nodeca/js-yaml/blob/master/CHANGELOG.md) - [Commits](https://github.com/nodeca/js-yaml/compare/4.1.0...4.1.1) --- updated-dependencies: - dependency-name: js-yaml dependency-version: 4.1.1 dependency-type: direct:production ... Signed-off-by: dependabot[bot] --- package-lock.json | 6 ++++-- package.json | 2 +- 2 files changed, 5 insertions(+), 3 deletions(-) diff --git a/package-lock.json b/package-lock.json index 28e08ed18..f1a3d294f 100644 --- a/package-lock.json +++ b/package-lock.json @@ -26,7 +26,7 @@ "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", "jsonschema": "1.4.1", "long": "^5.3.2", "node-forge": "^1.3.1", @@ -7089,7 +7089,9 @@ } }, "node_modules/js-yaml": { - "version": "4.1.0", + "version": "4.1.1", + "resolved": "https://registry.npmjs.org/js-yaml/-/js-yaml-4.1.1.tgz", + "integrity": "sha512-qQKT4zQxXl8lLwBtHMWwaTcGfFOZviOJet3Oy/xmGk2gZH677CJM9EvtfdSkgWcATZhj/55JZ0rmy3myCT5lsA==", "license": "MIT", "dependencies": { "argparse": "^2.0.1" diff --git a/package.json b/package.json index c7162d9c4..5a3378cea 100644 --- a/package.json +++ b/package.json @@ -41,7 +41,7 @@ "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", "jsonschema": "1.4.1", "long": "^5.3.2", "node-forge": "^1.3.1", From 8c254d05f319e87f0ef74fbae44c6146266462af Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Sat, 15 Nov 2025 10:57:22 +0000 Subject: [PATCH 16/51] Rebuild --- lib/analyze-action-post.js | 29 ++++++++++++++++------------- lib/analyze-action.js | 29 ++++++++++++++++------------- lib/autobuild-action.js | 29 ++++++++++++++++------------- lib/init-action-post.js | 29 ++++++++++++++++------------- lib/init-action.js | 29 ++++++++++++++++------------- lib/resolve-environment-action.js | 29 ++++++++++++++++------------- lib/setup-codeql-action.js | 29 ++++++++++++++++------------- lib/start-proxy-action-post.js | 29 ++++++++++++++++------------- lib/start-proxy-action.js | 29 ++++++++++++++++------------- lib/upload-lib.js | 29 ++++++++++++++++------------- lib/upload-sarif-action-post.js | 29 ++++++++++++++++------------- lib/upload-sarif-action.js | 29 ++++++++++++++++------------- 12 files changed, 192 insertions(+), 156 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index 5145c8bc2..7c1ecbeb0 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -27668,7 +27668,7 @@ var require_package = __commonJS({ "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", @@ -117368,6 +117368,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -117487,7 +117499,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -117527,16 +117539,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -120984,5 +120987,5 @@ tmp/lib/tmp.js: *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 4ee0d87e5..2a9d16c89 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -27668,7 +27668,7 @@ var require_package = __commonJS({ "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", @@ -85259,6 +85259,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -85378,7 +85390,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -85418,16 +85430,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -94159,7 +94162,7 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) long/index.js: (** diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index 8ff3dd6cd..de693821a 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -27668,7 +27668,7 @@ var require_package = __commonJS({ "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", @@ -81255,6 +81255,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -81374,7 +81386,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -81414,16 +81426,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -85456,5 +85459,5 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/init-action-post.js b/lib/init-action-post.js index af32a25c1..c48e9e142 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -27668,7 +27668,7 @@ var require_package = __commonJS({ "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", @@ -120266,6 +120266,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -120385,7 +120397,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -120425,16 +120437,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -128096,7 +128099,7 @@ tmp/lib/tmp.js: *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) long/index.js: (** diff --git a/lib/init-action.js b/lib/init-action.js index 6c92b83a7..2c35e5c68 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -27668,7 +27668,7 @@ var require_package = __commonJS({ "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", @@ -82564,6 +82564,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -82683,7 +82695,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -82723,16 +82735,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -90307,5 +90310,5 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index ff90d9c4e..4059c9db1 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -27668,7 +27668,7 @@ var require_package = __commonJS({ "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", @@ -81255,6 +81255,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -81374,7 +81386,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -81414,16 +81426,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -85077,5 +85080,5 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index 9cbb52ad3..3caba058d 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -27668,7 +27668,7 @@ var require_package = __commonJS({ "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", @@ -81311,6 +81311,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -81430,7 +81442,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -81470,16 +81482,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -86438,5 +86441,5 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index 2f1d4544a..3c7c44495 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -27668,7 +27668,7 @@ var require_package = __commonJS({ "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", @@ -117365,6 +117365,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -117484,7 +117496,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -117524,16 +117536,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -119875,5 +119878,5 @@ tmp/lib/tmp.js: *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index d51ba32df..cf0fbd92d 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -47326,7 +47326,7 @@ var require_package = __commonJS({ "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", @@ -97690,6 +97690,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -97809,7 +97821,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -97849,16 +97861,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -100580,5 +100583,5 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/upload-lib.js b/lib/upload-lib.js index fb897b861..dd1c9d395 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -28965,7 +28965,7 @@ var require_package = __commonJS({ "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", @@ -84180,6 +84180,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -84299,7 +84311,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -84339,16 +84351,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -90675,7 +90678,7 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) long/index.js: (** diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index 5a9cf35a2..abb2273e2 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -27668,7 +27668,7 @@ var require_package = __commonJS({ "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", @@ -117365,6 +117365,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -117484,7 +117496,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -117524,16 +117536,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -119932,5 +119935,5 @@ tmp/lib/tmp.js: *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index d0b2dd324..94002fa8f 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -27668,7 +27668,7 @@ var require_package = __commonJS({ "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", @@ -84153,6 +84153,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -84272,7 +84284,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -84312,16 +84324,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -91274,7 +91277,7 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) long/index.js: (** From ed3a01336ffabe8c1adaa9f89f7b2bb1667315d4 Mon Sep 17 00:00:00 2001 From: Mario Campos Date: Mon, 17 Nov 2025 08:56:34 -0600 Subject: [PATCH 17/51] Change v3 deprecation message to warning. --- lib/analyze-action-post.js | 968 ++++++++++++------------ lib/analyze-action.js | 714 +++++++++--------- lib/autobuild-action.js | 656 ++++++++-------- lib/init-action-post.js | 1002 ++++++++++++------------- lib/init-action.js | 706 ++++++++--------- lib/resolve-environment-action.js | 664 ++++++++-------- lib/setup-codeql-action.js | 668 ++++++++--------- lib/start-proxy-action-post.js | 940 +++++++++++------------ lib/start-proxy-action.js | 1168 ++++++++++++++--------------- lib/upload-lib.js | 648 ++++++++-------- lib/upload-sarif-action-post.js | 940 +++++++++++------------ lib/upload-sarif-action.js | 672 ++++++++--------- package-lock.json | 13 + src/util.test.ts | 2 +- src/util.ts | 2 +- 15 files changed, 4888 insertions(+), 4875 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index 7c1ecbeb0..d815da082 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug4("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay === "number" && delay > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve5(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug4; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning8(message, properties = {}) { + exports2.error = error3; + function warning9(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning8; + exports2.warning = warning9; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -27911,8 +27911,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27922,8 @@ var require_light = __commonJS({ return (await Promise.all(promises3)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28035,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28072,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28090,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } } doExpire(clearGlobalState, run, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28117,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28396,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve5, returned, task; + var args, cb, error3, reject, resolve5, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve5, reject } = this._queue.shift()); @@ -28407,9 +28407,9 @@ var require_light = __commonJS({ return resolve5(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28543,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28877,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29180,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29213,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29238,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -36085,8 +36085,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36320,9 +36320,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37653,14 +37653,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37670,7 +37670,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37678,8 +37678,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -37899,7 +37899,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39109,14 +39109,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39788,11 +39788,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39807,8 +39807,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40302,8 +40302,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40537,9 +40537,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41039,8 +41039,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41274,9 +41274,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42339,9 +42339,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42389,13 +42389,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42415,21 +42415,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42594,8 +42594,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43001,16 +43001,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43554,10 +43554,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45636,12 +45636,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -45906,16 +45906,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46040,7 +46040,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46206,7 +46206,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46447,9 +46447,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47157,7 +47157,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48220,26 +48220,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48280,12 +48280,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48293,13 +48293,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48308,7 +48308,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -62914,8 +62914,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -62949,10 +62949,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64536,8 +64536,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64552,9 +64552,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve5, reject) => { this.emitter.on("finish", resolve5); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65655,8 +65655,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65740,7 +65740,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -68892,7 +68892,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70259,9 +70259,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70375,12 +70375,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70414,13 +70414,13 @@ var require_requestUtils = __commonJS({ delay, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71236,8 +71236,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73083,8 +73083,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74655,12 +74655,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74680,12 +74680,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75149,8 +75149,8 @@ var require_test_transport = __commonJS({ } try { yield delay(this.responseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75161,8 +75161,8 @@ var require_test_transport = __commonJS({ stream.notifyMessage(msg); try { yield delay(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } } @@ -76225,8 +76225,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76322,8 +76322,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76354,18 +76354,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76633,8 +76633,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76835,22 +76835,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76905,15 +76905,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -76921,8 +76921,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76984,10 +76984,10 @@ var require_cache3 = __commonJS({ } core14.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else { @@ -77000,8 +77000,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77046,8 +77046,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core14.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core14.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError2(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core14.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77066,10 +77066,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77084,8 +77084,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79811,7 @@ var require_debug2 = __commonJS({ if (!debug4) { try { debug4 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug4 !== "function") { debug4 = function() { @@ -79844,8 +79844,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79919,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80088,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80290,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -83638,8 +83638,8 @@ var require_util11 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } function maskSecretUrls(body) { @@ -83732,8 +83732,8 @@ var require_artifact_twirp_client2 = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -83764,18 +83764,18 @@ var require_artifact_twirp_client2 = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -84046,11 +84046,11 @@ var require_blob_upload = __commonJS({ blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options), chunkTimer((0, config_1.getUploadChunkTimeout)()) ]); - } catch (error4) { - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + } catch (error3) { + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } - throw error4; + throw error3; } finally { abortController.abort(); } @@ -85071,9 +85071,9 @@ var require_async = __commonJS({ invokeCallback(callback, err && (err instanceof Error || err.message) ? err : new Error(err)); }); } - function invokeCallback(callback, error4, value) { + function invokeCallback(callback, error3, value) { try { - callback(error4, value); + callback(error3, value); } catch (err) { setImmediate$1((e) => { throw e; @@ -86379,10 +86379,10 @@ var require_async = __commonJS({ function reflect(fn) { var _fn = wrapAsync(fn); return initialParams(function reflectOn(args, reflectCallback) { - args.push((error4, ...cbArgs) => { + args.push((error3, ...cbArgs) => { let retVal = {}; - if (error4) { - retVal.error = error4; + if (error3) { + retVal.error = error3; } if (cbArgs.length > 0) { var value = cbArgs; @@ -86538,13 +86538,13 @@ var require_async = __commonJS({ var timer; function timeoutCallback() { var name = asyncFn.name || "anonymous"; - var error4 = new Error('Callback function "' + name + '" timed out.'); - error4.code = "ETIMEDOUT"; + var error3 = new Error('Callback function "' + name + '" timed out.'); + error3.code = "ETIMEDOUT"; if (info7) { - error4.info = info7; + error3.info = info7; } timedOut = true; - callback(error4); + callback(error3); } args.push((...cbArgs) => { if (!timedOut) { @@ -86587,7 +86587,7 @@ var require_async = __commonJS({ return callback[PROMISE_SYMBOL]; } function tryEach(tasks, callback) { - var error4 = null; + var error3 = null; var result; return eachSeries$1(tasks, (task, taskCb) => { wrapAsync(task)((err, ...args) => { @@ -86597,10 +86597,10 @@ var require_async = __commonJS({ } else { result = args; } - error4 = err; + error3 = err; taskCb(err ? null : {}); }); - }, () => callback(error4, result)); + }, () => callback(error3, result)); } var tryEach$1 = awaitify(tryEach); function unmemoize(fn) { @@ -93080,19 +93080,19 @@ var require_from = __commonJS({ next(); } }; - readable._destroy = function(error4, cb) { + readable._destroy = function(error3, cb) { PromisePrototypeThen( - close(error4), - () => process2.nextTick(cb, error4), + close(error3), + () => process2.nextTick(cb, error3), // nextTick is here in case cb throws - (e) => process2.nextTick(cb, e || error4) + (e) => process2.nextTick(cb, e || error3) ); }; - async function close(error4) { - const hadError = error4 !== void 0 && error4 !== null; + async function close(error3) { + const hadError = error3 !== void 0 && error3 !== null; const hasThrow = typeof iterator.throw === "function"; if (hadError && hasThrow) { - const { value, done } = await iterator.throw(error4); + const { value, done } = await iterator.throw(error3); await value; if (done) { return; @@ -93300,12 +93300,12 @@ var require_readable3 = __commonJS({ this.destroy(err); }; Readable.prototype[SymbolAsyncDispose] = function() { - let error4; + let error3; if (!this.destroyed) { - error4 = this.readableEnded ? null : new AbortError(); - this.destroy(error4); + error3 = this.readableEnded ? null : new AbortError(); + this.destroy(error3); } - return new Promise2((resolve5, reject) => eos(this, (err) => err && err !== error4 ? reject(err) : resolve5(null))); + return new Promise2((resolve5, reject) => eos(this, (err) => err && err !== error3 ? reject(err) : resolve5(null))); }; Readable.prototype.push = function(chunk, encoding) { return readableAddChunk(this, chunk, encoding, false); @@ -93858,14 +93858,14 @@ var require_readable3 = __commonJS({ } } stream.on("readable", next); - let error4; + let error3; const cleanup = eos( stream, { writable: false }, (err) => { - error4 = err ? aggregateTwoErrors(error4, err) : null; + error3 = err ? aggregateTwoErrors(error3, err) : null; callback(); callback = nop; } @@ -93875,19 +93875,19 @@ var require_readable3 = __commonJS({ const chunk = stream.destroyed ? null : stream.read(); if (chunk !== null) { yield chunk; - } else if (error4) { - throw error4; - } else if (error4 === null) { + } else if (error3) { + throw error3; + } else if (error3 === null) { return; } else { await new Promise2(next); } } } catch (err) { - error4 = aggregateTwoErrors(error4, err); - throw error4; + error3 = aggregateTwoErrors(error3, err); + throw error3; } finally { - if ((error4 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error4 === void 0 || stream._readableState.autoDestroy)) { + if ((error3 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error3 === void 0 || stream._readableState.autoDestroy)) { destroyImpl.destroyer(stream, null); } else { stream.off("readable", next); @@ -95380,11 +95380,11 @@ var require_pipeline3 = __commonJS({ yield* Readable.prototype[SymbolAsyncIterator].call(val2); } async function pumpToNode(iterable, writable, finish, { end }) { - let error4; + let error3; let onresolve = null; const resume = (err) => { if (err) { - error4 = err; + error3 = err; } if (onresolve) { const callback = onresolve; @@ -95393,12 +95393,12 @@ var require_pipeline3 = __commonJS({ } }; const wait = () => new Promise2((resolve5, reject) => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { onresolve = () => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve5(); } @@ -95428,7 +95428,7 @@ var require_pipeline3 = __commonJS({ } finish(); } catch (err) { - finish(error4 !== err ? aggregateTwoErrors(error4, err) : err); + finish(error3 !== err ? aggregateTwoErrors(error3, err) : err); } finally { cleanup(); writable.off("drain", resume); @@ -95482,7 +95482,7 @@ var require_pipeline3 = __commonJS({ if (outerSignal) { disposable = addAbortListener(outerSignal, abort); } - let error4; + let error3; let value; const destroys = []; let finishCount = 0; @@ -95491,23 +95491,23 @@ var require_pipeline3 = __commonJS({ } function finishImpl(err, final) { var _disposable; - if (err && (!error4 || error4.code === "ERR_STREAM_PREMATURE_CLOSE")) { - error4 = err; + if (err && (!error3 || error3.code === "ERR_STREAM_PREMATURE_CLOSE")) { + error3 = err; } - if (!error4 && !final) { + if (!error3 && !final) { return; } while (destroys.length) { - destroys.shift()(error4); + destroys.shift()(error3); } ; (_disposable = disposable) === null || _disposable === void 0 ? void 0 : _disposable[SymbolDispose](); ac.abort(); if (final) { - if (!error4) { + if (!error3) { lastStreamCleanup.forEach((fn) => fn()); } - process2.nextTick(callback, error4, value); + process2.nextTick(callback, error3, value); } } let ret; @@ -105849,18 +105849,18 @@ var require_zip_archive_output_stream = __commonJS({ ZipArchiveOutputStream.prototype._smartStream = function(ae, callback) { var deflate = ae.getMethod() === constants.METHOD_DEFLATED; var process2 = deflate ? new DeflateCRC32Stream(this.options.zlib) : new CRC32Stream(); - var error4 = null; + var error3 = null; function handleStuff() { var digest = process2.digest().readUInt32BE(0); ae.setCrc(digest); ae.setSize(process2.size()); ae.setCompressedSize(process2.size(true)); this._afterAppend(ae); - callback(error4, ae); + callback(error3, ae); } process2.once("end", handleStuff.bind(this)); process2.once("error", function(err) { - error4 = err; + error3 = err; }); process2.pipe(this, { end: false }); return process2; @@ -107226,11 +107226,11 @@ var require_streamx = __commonJS({ } [asyncIterator]() { const stream = this; - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("readable", onreadable); this.on("close", onclose); @@ -107262,7 +107262,7 @@ var require_streamx = __commonJS({ } function ondata(data) { if (promiseReject === null) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else if (data === null && (stream._duplexState & READ_DONE) === 0) promiseReject(STREAM_DESTROYED); else promiseResolve({ value: data, done: data === null }); promiseReject = promiseResolve = null; @@ -107436,7 +107436,7 @@ var require_streamx = __commonJS({ if (all.length < 2) throw new Error("Pipeline requires at least 2 streams"); let src = all[0]; let dest = null; - let error4 = null; + let error3 = null; for (let i = 1; i < all.length; i++) { dest = all[i]; if (isStreamx(src)) { @@ -107451,14 +107451,14 @@ var require_streamx = __commonJS({ let fin = false; const autoDestroy = isStreamx(dest) || !!(dest._writableState && dest._writableState.autoDestroy); dest.on("error", (err) => { - if (error4 === null) error4 = err; + if (error3 === null) error3 = err; }); dest.on("finish", () => { fin = true; - if (!autoDestroy) done(error4); + if (!autoDestroy) done(error3); }); if (autoDestroy) { - dest.on("close", () => done(error4 || (fin ? null : PREMATURE_CLOSE))); + dest.on("close", () => done(error3 || (fin ? null : PREMATURE_CLOSE))); } } return dest; @@ -107471,8 +107471,8 @@ var require_streamx = __commonJS({ } } function onerror(err) { - if (!err || error4) return; - error4 = err; + if (!err || error3) return; + error3 = err; for (const s of all) { s.destroy(err); } @@ -108038,7 +108038,7 @@ var require_extract = __commonJS({ cb(null); } [Symbol.asyncIterator]() { - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; let entryStream = null; @@ -108046,7 +108046,7 @@ var require_extract = __commonJS({ const extract2 = this; this.on("entry", onentry); this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("close", onclose); return { @@ -108070,8 +108070,8 @@ var require_extract = __commonJS({ cb(err); } function onnext(resolve5, reject) { - if (error4) { - return reject(error4); + if (error3) { + return reject(error3); } if (entryStream) { resolve5({ value: entryStream, done: false }); @@ -108097,9 +108097,9 @@ var require_extract = __commonJS({ } } function onclose() { - consumeCallback(error4); + consumeCallback(error3); if (!promiseResolve) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else promiseResolve({ value: void 0, done: true }); promiseResolve = promiseReject = null; } @@ -108995,18 +108995,18 @@ var require_zip2 = __commonJS({ return zipUploadStream; }); } - var zipErrorCallback = (error4) => { + var zipErrorCallback = (error3) => { core14.error("An error has occurred while creating the zip file for upload"); - core14.info(error4); + core14.info(error3); throw new Error("An error has occurred during zip creation for the artifact"); }; - var zipWarningCallback = (error4) => { - if (error4.code === "ENOENT") { + var zipWarningCallback = (error3) => { + if (error3.code === "ENOENT") { core14.warning("ENOENT warning during artifact zip creation. No such file or directory"); - core14.info(error4); + core14.info(error3); } else { - core14.warning(`A non-blocking warning has occurred during artifact zip creation: ${error4.code}`); - core14.info(error4); + core14.warning(`A non-blocking warning has occurred during artifact zip creation: ${error3.code}`); + core14.info(error3); } }; var zipFinishCallback = () => { @@ -110813,8 +110813,8 @@ var require_parser_stream = __commonJS({ this.unzipStream.on("entry", function(entry) { self2.push(entry); }); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(ParserStream, Transform); @@ -110954,8 +110954,8 @@ var require_extract2 = __commonJS({ this.createdDirectories = {}; var self2 = this; this.unzipStream.on("entry", this._processEntry.bind(this)); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(Extract, Transform); @@ -110989,8 +110989,8 @@ var require_extract2 = __commonJS({ self2.unfinishedEntries--; self2._notifyAwaiter(); }); - pipedStream.on("error", function(error4) { - self2.emit("error", error4); + pipedStream.on("error", function(error3) { + self2.emit("error", error3); }); entry.pipe(pipedStream); }; @@ -111125,11 +111125,11 @@ var require_download_artifact = __commonJS({ try { yield promises_1.default.access(path6); return true; - } catch (error4) { - if (error4.code === "ENOENT") { + } catch (error3) { + if (error3.code === "ENOENT") { return false; } else { - throw error4; + throw error3; } } }); @@ -111140,9 +111140,9 @@ var require_download_artifact = __commonJS({ while (retryCount < 5) { try { return yield streamExtractExternal(url, directory); - } catch (error4) { + } catch (error3) { retryCount++; - core14.debug(`Failed to download artifact after ${retryCount} retries due to ${error4.message}. Retrying in 5 seconds...`); + core14.debug(`Failed to download artifact after ${retryCount} retries due to ${error3.message}. Retrying in 5 seconds...`); yield new Promise((resolve5) => setTimeout(resolve5, 5e3)); } } @@ -111171,10 +111171,10 @@ var require_download_artifact = __commonJS({ const extractStream = passThrough; extractStream.on("data", () => { timer.refresh(); - }).on("error", (error4) => { - core14.debug(`response.message: Artifact download failed: ${error4.message}`); + }).on("error", (error3) => { + core14.debug(`response.message: Artifact download failed: ${error3.message}`); clearTimeout(timer); - reject(error4); + reject(error3); }).pipe(unzip_stream_1.default.Extract({ path: directory })).on("close", () => { clearTimeout(timer); if (hashStream) { @@ -111183,8 +111183,8 @@ var require_download_artifact = __commonJS({ core14.info(`SHA256 digest of downloaded artifact is ${sha256Digest}`); } resolve5({ sha256Digest: `sha256:${sha256Digest}` }); - }).on("error", (error4) => { - reject(error4); + }).on("error", (error3) => { + reject(error3); }); }); }); @@ -111223,8 +111223,8 @@ var require_download_artifact = __commonJS({ core14.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -111266,8 +111266,8 @@ Are you trying to download from a different run? Try specifying a github-token w core14.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -111365,9 +111365,9 @@ var require_dist_node16 = __commonJS({ return request(options).then((response) => { octokit.log.info(`${requestOptions.method} ${path6} - ${response.status} in ${Date.now() - start}ms`); return response; - }).catch((error4) => { - octokit.log.info(`${requestOptions.method} ${path6} - ${error4.status} in ${Date.now() - start}ms`); - throw error4; + }).catch((error3) => { + octokit.log.info(`${requestOptions.method} ${path6} - ${error3.status} in ${Date.now() - start}ms`); + throw error3; }); }); } @@ -111385,22 +111385,22 @@ var require_dist_node17 = __commonJS({ return ex && typeof ex === "object" && "default" in ex ? ex["default"] : ex; } var Bottleneck = _interopDefault(require_light()); - async function errorRequest(octokit, state, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(octokit, state, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } async function wrapRequest(state, request, options) { const limiter = new Bottleneck(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -111422,12 +111422,12 @@ var require_dist_node17 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -111918,13 +111918,13 @@ var require_client2 = __commonJS({ throw new errors_1.GHESNotSupportedError(); } return (0, upload_artifact_1.uploadArtifact)(name, files, rootDirectory, options); - } catch (error4) { - (0, core_1.warning)(`Artifact upload failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Artifact upload failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions is operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -111939,13 +111939,13 @@ If the error persists, please check whether Actions is operating normally at [ht return (0, download_artifact_1.downloadArtifactPublic)(artifactId, repositoryOwner, repositoryName, token, downloadOptions); } return (0, download_artifact_1.downloadArtifactInternal)(artifactId, options); - } catch (error4) { - (0, core_1.warning)(`Download Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Download Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -111960,13 +111960,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, list_artifacts_1.listArtifactsPublic)(workflowRunId, repositoryOwner, repositoryName, token, options === null || options === void 0 ? void 0 : options.latest); } return (0, list_artifacts_1.listArtifactsInternal)(options === null || options === void 0 ? void 0 : options.latest); - } catch (error4) { - (0, core_1.warning)(`Listing Artifacts failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Listing Artifacts failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -111981,13 +111981,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, get_artifact_1.getArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, get_artifact_1.getArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Get Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Get Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -112002,13 +112002,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, delete_artifact_1.deleteArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, delete_artifact_1.deleteArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Delete Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Delete Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -112508,14 +112508,14 @@ var require_tmp = __commonJS({ options.template = _getRelativePathSync("template", options.template, tmpDir); return options; } - function _isEBADF(error4) { - return _isExpectedError(error4, -EBADF, "EBADF"); + function _isEBADF(error3) { + return _isExpectedError(error3, -EBADF, "EBADF"); } - function _isENOENT(error4) { - return _isExpectedError(error4, -ENOENT, "ENOENT"); + function _isENOENT(error3) { + return _isExpectedError(error3, -ENOENT, "ENOENT"); } - function _isExpectedError(error4, errno, code) { - return IS_WIN32 ? error4.code === code : error4.code === code && error4.errno === errno; + function _isExpectedError(error3, errno, code) { + return IS_WIN32 ? error3.code === code : error3.code === code && error3.errno === errno; } function setGracefulCleanup() { _gracefulCleanup = true; @@ -114227,9 +114227,9 @@ var require_upload_gzip = __commonJS({ const size = (yield stat(tempFilePath)).size; resolve5(size); })); - outputStream.on("error", (error4) => { - console.log(error4); - reject(error4); + outputStream.on("error", (error3) => { + console.log(error3); + reject(error3); }); }); }); @@ -114354,9 +114354,9 @@ var require_requestUtils2 = __commonJS({ } isRetryable = (0, utils_1.isRetryableStatusCode)(statusCode); errorMessage = `Artifact service responded with ${statusCode}`; - } catch (error4) { + } catch (error3) { isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { core14.info(`${name} - Error is not retryable`); @@ -114722,9 +114722,9 @@ var require_upload_http_client = __commonJS({ let response; try { response = yield uploadChunkRequest(); - } catch (error4) { + } catch (error3) { core14.info(`An error has been caught http-client index ${httpClientIndex}, retrying the upload`); - console.log(error4); + console.log(error3); if (incrementAndCheckRetryLimit()) { return false; } @@ -114913,8 +114913,8 @@ var require_download_http_client = __commonJS({ } this.statusReporter.incrementProcessedCount(); } - }))).catch((error4) => { - throw new Error(`Unable to download the artifact: ${error4}`); + }))).catch((error3) => { + throw new Error(`Unable to download the artifact: ${error3}`); }).finally(() => { this.statusReporter.stop(); this.downloadHttpManager.disposeAndReplaceAllClients(); @@ -114979,9 +114979,9 @@ var require_download_http_client = __commonJS({ let response; try { response = yield makeDownloadRequest(); - } catch (error4) { + } catch (error3) { core14.info("An error occurred while attempting to download a file"); - console.log(error4); + console.log(error3); yield backOff(); continue; } @@ -114995,7 +114995,7 @@ var require_download_http_client = __commonJS({ } else { forceRetry = true; } - } catch (error4) { + } catch (error3) { forceRetry = true; } } @@ -115021,31 +115021,31 @@ var require_download_http_client = __commonJS({ yield new Promise((resolve5, reject) => { if (isGzip) { const gunzip = zlib.createGunzip(); - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core14.info(`An error occurred while attempting to read the response stream`); gunzip.close(); destinationStream.close(); - reject(error4); - }).pipe(gunzip).on("error", (error4) => { + reject(error3); + }).pipe(gunzip).on("error", (error3) => { core14.info(`An error occurred while attempting to decompress the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve5(); - }).on("error", (error4) => { + }).on("error", (error3) => { core14.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } else { - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core14.info(`An error occurred while attempting to read the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve5(); - }).on("error", (error4) => { + }).on("error", (error3) => { core14.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } }); @@ -116464,14 +116464,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs7.lstat(itemPath, { bigint: true }) : await fs7.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs7.lstat(itemPath, { bigint: true }) : await fs7.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs7.readdir(itemPath) : await fs7.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs7.readdir(itemPath) : await fs7.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -116482,13 +116482,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -119109,9 +119109,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -119204,11 +119204,11 @@ function getCachedCodeQlVersion() { async function codeQlVersionAtLeast(codeql, requiredVersion) { return semver.gte((await codeql.getVersion()).version, requiredVersion); } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } function cloneObject(obj) { return JSON.parse(JSON.stringify(obj)); @@ -119409,19 +119409,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -119437,9 +119437,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -119674,13 +119674,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -120890,15 +120890,15 @@ async function runWrapper() { if (fs6.existsSync(javaTempDependencyDir)) { try { fs6.rmSync(javaTempDependencyDir, { recursive: true }); - } catch (error4) { + } catch (error3) { logger.info( - `Failed to remove temporary Java dependencies directory: ${getErrorMessage(error4)}` + `Failed to remove temporary Java dependencies directory: ${getErrorMessage(error3)}` ); } } - } catch (error4) { + } catch (error3) { core13.setFailed( - `analyze post-action step failed: ${getErrorMessage(error4)}` + `analyze post-action step failed: ${getErrorMessage(error3)}` ); } } diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 2a9d16c89..5dcfd5062 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug5("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve8(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug5; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning9(message, properties = {}) { + exports2.error = error3; + function warning10(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning9; + exports2.warning = warning10; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url2 = ""; return { value: { @@ -27911,8 +27911,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27922,8 @@ var require_light = __commonJS({ return (await Promise.all(promises4)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28035,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28072,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run2, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28090,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } } doExpire(clearGlobalState, run2, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run2, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run2, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28117,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28396,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve8, returned, task; + var args, cb, error3, reject, resolve8, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve8, reject } = this._queue.shift()); @@ -28407,9 +28407,9 @@ var require_light = __commonJS({ return resolve8(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28543,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28877,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29180,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29213,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29238,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -36085,8 +36085,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36320,9 +36320,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37653,14 +37653,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37670,7 +37670,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37678,8 +37678,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -37899,7 +37899,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39109,14 +39109,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39788,11 +39788,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39807,8 +39807,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40302,8 +40302,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40537,9 +40537,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41039,8 +41039,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41274,9 +41274,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42339,9 +42339,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42389,13 +42389,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42415,21 +42415,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42594,8 +42594,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43001,16 +43001,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43554,10 +43554,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45636,12 +45636,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -45906,16 +45906,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46040,7 +46040,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46206,7 +46206,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46447,9 +46447,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47157,7 +47157,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48220,26 +48220,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48280,12 +48280,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48293,13 +48293,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48308,7 +48308,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -62914,8 +62914,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -62949,10 +62949,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64536,8 +64536,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64552,9 +64552,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve8, reject) => { this.emitter.on("finish", resolve8); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65655,8 +65655,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65740,7 +65740,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -68892,7 +68892,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70259,9 +70259,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core15.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core15.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70375,12 +70375,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70414,13 +70414,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71236,8 +71236,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73083,8 +73083,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74655,12 +74655,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74680,12 +74680,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75149,8 +75149,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75161,8 +75161,8 @@ var require_test_transport = __commonJS({ stream2.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } } @@ -76225,8 +76225,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url2} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url2} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76322,8 +76322,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url2, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76354,18 +76354,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76633,8 +76633,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76835,22 +76835,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core15.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core15.error(`Failed to restore: ${error4.message}`); + core15.error(`Failed to restore: ${error3.message}`); } else { - core15.warning(`Failed to restore: ${error4.message}`); + core15.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core15.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core15.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76905,15 +76905,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core15.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core15.error(`Failed to restore: ${error4.message}`); + core15.error(`Failed to restore: ${error3.message}`); } else { - core15.warning(`Failed to restore: ${error4.message}`); + core15.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -76921,8 +76921,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core15.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core15.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76984,10 +76984,10 @@ var require_cache3 = __commonJS({ } core15.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core15.info(`Failed to save: ${typedError.message}`); } else { @@ -77000,8 +77000,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core15.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core15.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77046,8 +77046,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core15.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core15.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError2(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core15.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77066,10 +77066,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core15.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77084,8 +77084,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core15.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core15.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79811,7 @@ var require_debug2 = __commonJS({ if (!debug5) { try { debug5 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug5 !== "function") { debug5 = function() { @@ -79844,8 +79844,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79919,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80088,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80290,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -84355,14 +84355,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs17.lstat(itemPath, { bigint: true }) : await fs17.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs17.lstat(itemPath, { bigint: true }) : await fs17.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs17.readdir(itemPath) : await fs17.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs17.readdir(itemPath) : await fs17.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -84373,13 +84373,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -86996,9 +86996,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -87382,11 +87382,11 @@ function parseMatrixInput(matrixInput) { } return JSON.parse(matrixInput); } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } async function checkDiskUsage(logger) { try { @@ -87409,9 +87409,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -87423,7 +87423,7 @@ function checkActionVersion(version, githubVersion) { semver.coerce(githubVersion.version) ?? "0.0.0", ">=3.20" )) { - core3.error( + core3.warning( "CodeQL Action v3 will be deprecated in December 2026. Please update all occurrences of the CodeQL Action in your workflow files to v4. For more information, see https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/" ); core3.exportVariable("CODEQL_ACTION_DID_LOG_VERSION_DEPRECATION" /* LOG_VERSION_DEPRECATION */, "true"); @@ -87961,19 +87961,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -87989,9 +87989,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -88243,13 +88243,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -88592,9 +88592,9 @@ async function uploadOverlayBaseDatabaseToCache(codeql, config, logger) { logger.warning("Timed out while uploading overlay-base database"); return false; } - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to upload overlay-base database to cache: ${error4 instanceof Error ? error4.message : String(error4)}` + `Failed to upload overlay-base database to cache: ${error3 instanceof Error ? error3.message : String(error3)}` ); return false; } @@ -89176,17 +89176,17 @@ async function getFileDiffsWithBasehead(branches, logger) { ${JSON.stringify(response, null, 2)}` ); return response.data.files; - } catch (error4) { - if (error4.status) { - logger.warning(`Error retrieving diff ${basehead}: ${error4.message}`); + } catch (error3) { + if (error3.status) { + logger.warning(`Error retrieving diff ${basehead}: ${error3.message}`); logger.debug( `Error running compareCommitsWithBasehead(${basehead}): -Request: ${JSON.stringify(error4.request, null, 2)} -Error Response: ${JSON.stringify(error4.response, null, 2)}` +Request: ${JSON.stringify(error3.request, null, 2)} +Error Response: ${JSON.stringify(error3.response, null, 2)}` ); return void 0; } else { - throw error4; + throw error3; } } } @@ -91184,15 +91184,15 @@ async function uploadDependencyCaches(codeql, features, config, logger) { upload_size_bytes: Math.round(size), upload_duration_ms }); - } catch (error4) { - if (error4 instanceof actionsCache3.ReserveCacheError) { + } catch (error3) { + if (error3 instanceof actionsCache3.ReserveCacheError) { logger.info( `Not uploading cache for ${language}, because ${key} is already in use.` ); - logger.debug(error4.message); + logger.debug(error3.message); status.push({ language, result: "duplicate" /* Duplicate */ }); } else { - throw error4; + throw error3; } } } @@ -91290,11 +91290,11 @@ function writeDiagnostic(config, language, diagnostic) { // src/analyze.ts var CodeQLAnalysisError = class extends Error { - constructor(queriesStatusReport, message, error4) { + constructor(queriesStatusReport, message, error3) { super(message); this.queriesStatusReport = queriesStatusReport; this.message = message; - this.error = error4; + this.error = error3; this.name = "CodeQLAnalysisError"; } }; @@ -91613,9 +91613,9 @@ async function runQueries(sarifFolder, memoryFlag, threadsFlag, diffRangePackDir async function runFinalize(outputDir, threadsFlag, memoryFlag, codeql, config, logger) { try { await fs12.promises.rm(outputDir, { force: true, recursive: true }); - } catch (error4) { - if (error4?.code !== "ENOENT") { - throw error4; + } catch (error3) { + if (error3?.code !== "ENOENT") { + throw error3; } } await fs12.promises.mkdir(outputDir, { recursive: true }); @@ -91740,9 +91740,9 @@ function isFirstPartyAnalysis(actionName) { } return process.env["CODEQL_ACTION_INIT_HAS_RUN" /* INIT_ACTION_HAS_RUN */] === "true"; } -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -93409,15 +93409,15 @@ function validateSarifFileSchema(sarif, sarifFilePath, logger) { const warnings = (result.errors ?? []).filter( (err) => err.name === "format" && typeof err.argument === "string" && warningAttributes.includes(err.argument) ); - for (const warning9 of warnings) { + for (const warning10 of warnings) { logger.info( - `Warning: '${warning9.instance}' is not a valid URI in '${warning9.property}'.` + `Warning: '${warning10.instance}' is not a valid URI in '${warning10.property}'.` ); } if (errors.length > 0) { - for (const error4 of errors) { - logger.startGroup(`Error details: ${error4.stack}`); - logger.info(JSON.stringify(error4, null, 2)); + for (const error3 of errors) { + logger.startGroup(`Error details: ${error3.stack}`); + logger.info(JSON.stringify(error3, null, 2)); logger.endGroup(); } const sarifErrors = errors.map((e) => `- ${e.stack}`); @@ -93670,10 +93670,10 @@ function shouldConsiderConfigurationError(processingErrors) { } function shouldConsiderInvalidRequest(processingErrors) { return processingErrors.every( - (error4) => error4.startsWith("rejecting SARIF") || error4.startsWith("an invalid URI was provided as a SARIF location") || error4.startsWith("locationFromSarifResult: expected artifact location") || error4.startsWith( + (error3) => error3.startsWith("rejecting SARIF") || error3.startsWith("an invalid URI was provided as a SARIF location") || error3.startsWith("locationFromSarifResult: expected artifact location") || error3.startsWith( "could not convert rules: invalid security severity value, is not a number" ) || /^SARIF URI scheme [^\s]* did not match the checkout URI scheme [^\s]*/.test( - error4 + error3 ) ); } @@ -93789,8 +93789,8 @@ async function postProcessAndUploadSarif(logger, features, uploadKind, checkoutP } // src/analyze-action.ts -async function sendStatusReport2(startedAt, config, stats, error4, trapCacheUploadTime, dbCreationTimings, didUploadTrapCaches, trapCacheCleanup, dependencyCacheResults, logger) { - const status = getActionsStatus(error4, stats?.analyze_failure_language); +async function sendStatusReport2(startedAt, config, stats, error3, trapCacheUploadTime, dbCreationTimings, didUploadTrapCaches, trapCacheCleanup, dependencyCacheResults, logger) { + const status = getActionsStatus(error3, stats?.analyze_failure_language); const statusReportBase = await createStatusReportBase( "finish" /* Analyze */, status, @@ -93798,8 +93798,8 @@ async function sendStatusReport2(startedAt, config, stats, error4, trapCacheUplo config, await checkDiskUsage(logger), logger, - error4?.message, - error4?.stack + error3?.message, + error3?.stack ); if (statusReportBase !== void 0) { const report = { @@ -94077,15 +94077,15 @@ async function run() { } core14.exportVariable("CODEQL_ACTION_ANALYZE_DID_COMPLETE_SUCCESSFULLY" /* ANALYZE_DID_COMPLETE_SUCCESSFULLY */, "true"); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); if (getOptionalInput("expect-error") !== "true" || hasBadExpectErrorInput()) { - core14.setFailed(error4.message); + core14.setFailed(error3.message); } await sendStatusReport2( startedAt, config, - error4 instanceof CodeQLAnalysisError ? error4.queriesStatusReport : void 0, - error4 instanceof CodeQLAnalysisError ? error4.error : error4, + error3 instanceof CodeQLAnalysisError ? error3.queriesStatusReport : void 0, + error3 instanceof CodeQLAnalysisError ? error3.error : error3, trapCacheUploadTime, dbCreationTimings, didUploadTrapCaches, @@ -94143,8 +94143,8 @@ var runPromise = run(); async function runWrapper() { try { await runPromise; - } catch (error4) { - core14.setFailed(`analyze action failed: ${getErrorMessage(error4)}`); + } catch (error3) { + core14.setFailed(`analyze action failed: ${getErrorMessage(error3)}`); } await checkForTimeout(); } diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index de693821a..2655e4315 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug5("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay === "number" && delay > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve5(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug5; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning8(message, properties = {}) { + exports2.error = error3; + function warning9(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning8; + exports2.warning = warning9; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -27911,8 +27911,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27922,8 @@ var require_light = __commonJS({ return (await Promise.all(promises2)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28035,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28072,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run2, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28090,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } } doExpire(clearGlobalState, run2, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run2, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run2, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28117,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28396,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve5, returned, task; + var args, cb, error3, reject, resolve5, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve5, reject } = this._queue.shift()); @@ -28407,9 +28407,9 @@ var require_light = __commonJS({ return resolve5(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28543,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28877,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29180,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29213,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29238,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -36085,8 +36085,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36320,9 +36320,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37653,14 +37653,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37670,7 +37670,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37678,8 +37678,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -37899,7 +37899,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39109,14 +39109,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39788,11 +39788,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39807,8 +39807,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40302,8 +40302,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40537,9 +40537,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41039,8 +41039,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41274,9 +41274,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42339,9 +42339,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42389,13 +42389,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42415,21 +42415,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42594,8 +42594,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43001,16 +43001,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43554,10 +43554,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45636,12 +45636,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -45906,16 +45906,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46040,7 +46040,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46206,7 +46206,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46447,9 +46447,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47157,7 +47157,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48220,26 +48220,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48280,12 +48280,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48293,13 +48293,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48308,7 +48308,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -62914,8 +62914,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -62949,10 +62949,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64536,8 +64536,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64552,9 +64552,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve5, reject) => { this.emitter.on("finish", resolve5); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65655,8 +65655,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65740,7 +65740,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -68892,7 +68892,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70259,9 +70259,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70375,12 +70375,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70414,13 +70414,13 @@ var require_requestUtils = __commonJS({ delay, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71236,8 +71236,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73083,8 +73083,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74655,12 +74655,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74680,12 +74680,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75149,8 +75149,8 @@ var require_test_transport = __commonJS({ } try { yield delay(this.responseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75161,8 +75161,8 @@ var require_test_transport = __commonJS({ stream.notifyMessage(msg); try { yield delay(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } } @@ -76225,8 +76225,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76322,8 +76322,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76354,18 +76354,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76633,8 +76633,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76835,22 +76835,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76905,15 +76905,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -76921,8 +76921,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76984,10 +76984,10 @@ var require_cache3 = __commonJS({ } core14.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core14.info(`Failed to save: ${typedError.message}`); } else { @@ -77000,8 +77000,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77046,8 +77046,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core14.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core14.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core14.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77066,10 +77066,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core14.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77084,8 +77084,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79811,7 @@ var require_debug2 = __commonJS({ if (!debug5) { try { debug5 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug5 !== "function") { debug5 = function() { @@ -79844,8 +79844,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79919,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80088,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80290,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -80351,14 +80351,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs7.lstat(itemPath, { bigint: true }) : await fs7.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs7.lstat(itemPath, { bigint: true }) : await fs7.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs7.readdir(itemPath) : await fs7.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs7.readdir(itemPath) : await fs7.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -80369,13 +80369,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -82996,9 +82996,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -83126,11 +83126,11 @@ function getTestingEnvironment() { } return testingEnvironment; } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } async function checkDiskUsage(logger) { try { @@ -83153,9 +83153,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -83167,7 +83167,7 @@ function checkActionVersion(version, githubVersion) { semver.coerce(githubVersion.version) ?? "0.0.0", ">=3.20" )) { - core3.error( + core3.warning( "CodeQL Action v3 will be deprecated in December 2026. Please update all occurrences of the CodeQL Action in your workflow files to v4. For more information, see https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/" ); core3.exportVariable("CODEQL_ACTION_DID_LOG_VERSION_DEPRECATION" /* LOG_VERSION_DEPRECATION */, "true"); @@ -83462,19 +83462,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -83490,9 +83490,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -83733,13 +83733,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -85186,9 +85186,9 @@ function isFirstPartyAnalysis(actionName) { } return process.env["CODEQL_ACTION_INIT_HAS_RUN" /* INIT_ACTION_HAS_RUN */] === "true"; } -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -85425,9 +85425,9 @@ async function run() { } await endTracingForCluster(codeql, config, logger); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); core13.setFailed( - `We were unable to automatically build your code. Please replace the call to the autobuild action with your custom build steps. ${error4.message}` + `We were unable to automatically build your code. Please replace the call to the autobuild action with your custom build steps. ${error3.message}` ); await sendCompletedStatusReport( config, @@ -85435,7 +85435,7 @@ async function run() { startedAt, languages ?? [], currentLanguage, - error4 + error3 ); return; } @@ -85445,8 +85445,8 @@ async function run() { async function runWrapper() { try { await run(); - } catch (error4) { - core13.setFailed(`autobuild action failed. ${getErrorMessage(error4)}`); + } catch (error3) { + core13.setFailed(`autobuild action failed. ${getErrorMessage(error3)}`); } } void runWrapper(); diff --git a/lib/init-action-post.js b/lib/init-action-post.js index c48e9e142..c238f7d06 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug5("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve8(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug5; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning11(message, properties = {}) { + exports2.error = error3; + function warning12(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning11; + exports2.warning = warning12; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url2 = ""; return { value: { @@ -27911,8 +27911,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27922,8 @@ var require_light = __commonJS({ return (await Promise.all(promises5)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28035,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28072,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run2, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28090,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } } doExpire(clearGlobalState, run2, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run2, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run2, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28117,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28396,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve8, returned, task; + var args, cb, error3, reject, resolve8, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve8, reject } = this._queue.shift()); @@ -28407,9 +28407,9 @@ var require_light = __commonJS({ return resolve8(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28543,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28877,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29180,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29213,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29238,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -36085,8 +36085,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36320,9 +36320,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37653,14 +37653,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37670,7 +37670,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37678,8 +37678,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -37899,7 +37899,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39109,14 +39109,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39788,11 +39788,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39807,8 +39807,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40302,8 +40302,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40537,9 +40537,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41039,8 +41039,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41274,9 +41274,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42339,9 +42339,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42389,13 +42389,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42415,21 +42415,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42594,8 +42594,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43001,16 +43001,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43554,10 +43554,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45636,12 +45636,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -45906,16 +45906,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46040,7 +46040,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46206,7 +46206,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46447,9 +46447,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47157,7 +47157,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48220,26 +48220,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48280,12 +48280,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48293,13 +48293,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48308,7 +48308,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -62914,8 +62914,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -62949,10 +62949,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64536,8 +64536,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64552,9 +64552,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve8, reject) => { this.emitter.on("finish", resolve8); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65655,8 +65655,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65740,7 +65740,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -68892,7 +68892,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70259,9 +70259,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core18.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core18.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70375,12 +70375,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70414,13 +70414,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71236,8 +71236,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73083,8 +73083,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74655,12 +74655,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74680,12 +74680,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75149,8 +75149,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75161,8 +75161,8 @@ var require_test_transport = __commonJS({ stream2.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } } @@ -76225,8 +76225,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url2} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url2} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76322,8 +76322,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url2, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76354,18 +76354,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76633,8 +76633,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76835,22 +76835,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core18.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core18.error(`Failed to restore: ${error4.message}`); + core18.error(`Failed to restore: ${error3.message}`); } else { - core18.warning(`Failed to restore: ${error4.message}`); + core18.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core18.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core18.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76905,15 +76905,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core18.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core18.error(`Failed to restore: ${error4.message}`); + core18.error(`Failed to restore: ${error3.message}`); } else { - core18.warning(`Failed to restore: ${error4.message}`); + core18.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -76921,8 +76921,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core18.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core18.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76984,10 +76984,10 @@ var require_cache3 = __commonJS({ } core18.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core18.info(`Failed to save: ${typedError.message}`); } else { @@ -77000,8 +77000,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core18.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core18.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77046,8 +77046,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core18.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core18.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError2(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core18.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77066,10 +77066,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core18.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77084,8 +77084,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core18.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core18.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79811,7 @@ var require_debug2 = __commonJS({ if (!debug5) { try { debug5 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug5 !== "function") { debug5 = function() { @@ -79844,8 +79844,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79919,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80088,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80290,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -83638,8 +83638,8 @@ var require_util11 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url2} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url2} ${error3 instanceof Error ? error3.message : String(error3)}`); } } function maskSecretUrls(body) { @@ -83732,8 +83732,8 @@ var require_artifact_twirp_client2 = __commonJS({ return this.httpClient.post(url2, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -83764,18 +83764,18 @@ var require_artifact_twirp_client2 = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -84046,11 +84046,11 @@ var require_blob_upload = __commonJS({ blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options), chunkTimer((0, config_1.getUploadChunkTimeout)()) ]); - } catch (error4) { - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + } catch (error3) { + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } - throw error4; + throw error3; } finally { abortController.abort(); } @@ -85071,9 +85071,9 @@ var require_async = __commonJS({ invokeCallback(callback, err && (err instanceof Error || err.message) ? err : new Error(err)); }); } - function invokeCallback(callback, error4, value) { + function invokeCallback(callback, error3, value) { try { - callback(error4, value); + callback(error3, value); } catch (err) { setImmediate$1((e) => { throw e; @@ -86379,10 +86379,10 @@ var require_async = __commonJS({ function reflect(fn) { var _fn = wrapAsync(fn); return initialParams(function reflectOn(args, reflectCallback) { - args.push((error4, ...cbArgs) => { + args.push((error3, ...cbArgs) => { let retVal = {}; - if (error4) { - retVal.error = error4; + if (error3) { + retVal.error = error3; } if (cbArgs.length > 0) { var value = cbArgs; @@ -86538,13 +86538,13 @@ var require_async = __commonJS({ var timer; function timeoutCallback() { var name = asyncFn.name || "anonymous"; - var error4 = new Error('Callback function "' + name + '" timed out.'); - error4.code = "ETIMEDOUT"; + var error3 = new Error('Callback function "' + name + '" timed out.'); + error3.code = "ETIMEDOUT"; if (info7) { - error4.info = info7; + error3.info = info7; } timedOut = true; - callback(error4); + callback(error3); } args.push((...cbArgs) => { if (!timedOut) { @@ -86587,7 +86587,7 @@ var require_async = __commonJS({ return callback[PROMISE_SYMBOL]; } function tryEach(tasks, callback) { - var error4 = null; + var error3 = null; var result; return eachSeries$1(tasks, (task, taskCb) => { wrapAsync(task)((err, ...args) => { @@ -86597,10 +86597,10 @@ var require_async = __commonJS({ } else { result = args; } - error4 = err; + error3 = err; taskCb(err ? null : {}); }); - }, () => callback(error4, result)); + }, () => callback(error3, result)); } var tryEach$1 = awaitify(tryEach); function unmemoize(fn) { @@ -93080,19 +93080,19 @@ var require_from = __commonJS({ next(); } }; - readable._destroy = function(error4, cb) { + readable._destroy = function(error3, cb) { PromisePrototypeThen( - close(error4), - () => process2.nextTick(cb, error4), + close(error3), + () => process2.nextTick(cb, error3), // nextTick is here in case cb throws - (e) => process2.nextTick(cb, e || error4) + (e) => process2.nextTick(cb, e || error3) ); }; - async function close(error4) { - const hadError = error4 !== void 0 && error4 !== null; + async function close(error3) { + const hadError = error3 !== void 0 && error3 !== null; const hasThrow = typeof iterator.throw === "function"; if (hadError && hasThrow) { - const { value, done } = await iterator.throw(error4); + const { value, done } = await iterator.throw(error3); await value; if (done) { return; @@ -93300,12 +93300,12 @@ var require_readable3 = __commonJS({ this.destroy(err); }; Readable2.prototype[SymbolAsyncDispose] = function() { - let error4; + let error3; if (!this.destroyed) { - error4 = this.readableEnded ? null : new AbortError(); - this.destroy(error4); + error3 = this.readableEnded ? null : new AbortError(); + this.destroy(error3); } - return new Promise2((resolve8, reject) => eos(this, (err) => err && err !== error4 ? reject(err) : resolve8(null))); + return new Promise2((resolve8, reject) => eos(this, (err) => err && err !== error3 ? reject(err) : resolve8(null))); }; Readable2.prototype.push = function(chunk, encoding) { return readableAddChunk(this, chunk, encoding, false); @@ -93858,14 +93858,14 @@ var require_readable3 = __commonJS({ } } stream2.on("readable", next); - let error4; + let error3; const cleanup = eos( stream2, { writable: false }, (err) => { - error4 = err ? aggregateTwoErrors(error4, err) : null; + error3 = err ? aggregateTwoErrors(error3, err) : null; callback(); callback = nop; } @@ -93875,19 +93875,19 @@ var require_readable3 = __commonJS({ const chunk = stream2.destroyed ? null : stream2.read(); if (chunk !== null) { yield chunk; - } else if (error4) { - throw error4; - } else if (error4 === null) { + } else if (error3) { + throw error3; + } else if (error3 === null) { return; } else { await new Promise2(next); } } } catch (err) { - error4 = aggregateTwoErrors(error4, err); - throw error4; + error3 = aggregateTwoErrors(error3, err); + throw error3; } finally { - if ((error4 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error4 === void 0 || stream2._readableState.autoDestroy)) { + if ((error3 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error3 === void 0 || stream2._readableState.autoDestroy)) { destroyImpl.destroyer(stream2, null); } else { stream2.off("readable", next); @@ -95380,11 +95380,11 @@ var require_pipeline3 = __commonJS({ yield* Readable2.prototype[SymbolAsyncIterator].call(val2); } async function pumpToNode(iterable, writable, finish, { end }) { - let error4; + let error3; let onresolve = null; const resume = (err) => { if (err) { - error4 = err; + error3 = err; } if (onresolve) { const callback = onresolve; @@ -95393,12 +95393,12 @@ var require_pipeline3 = __commonJS({ } }; const wait = () => new Promise2((resolve8, reject) => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { onresolve = () => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve8(); } @@ -95428,7 +95428,7 @@ var require_pipeline3 = __commonJS({ } finish(); } catch (err) { - finish(error4 !== err ? aggregateTwoErrors(error4, err) : err); + finish(error3 !== err ? aggregateTwoErrors(error3, err) : err); } finally { cleanup(); writable.off("drain", resume); @@ -95482,7 +95482,7 @@ var require_pipeline3 = __commonJS({ if (outerSignal) { disposable = addAbortListener(outerSignal, abort); } - let error4; + let error3; let value; const destroys = []; let finishCount = 0; @@ -95491,23 +95491,23 @@ var require_pipeline3 = __commonJS({ } function finishImpl(err, final) { var _disposable; - if (err && (!error4 || error4.code === "ERR_STREAM_PREMATURE_CLOSE")) { - error4 = err; + if (err && (!error3 || error3.code === "ERR_STREAM_PREMATURE_CLOSE")) { + error3 = err; } - if (!error4 && !final) { + if (!error3 && !final) { return; } while (destroys.length) { - destroys.shift()(error4); + destroys.shift()(error3); } ; (_disposable = disposable) === null || _disposable === void 0 ? void 0 : _disposable[SymbolDispose](); ac.abort(); if (final) { - if (!error4) { + if (!error3) { lastStreamCleanup.forEach((fn) => fn()); } - process2.nextTick(callback, error4, value); + process2.nextTick(callback, error3, value); } } let ret; @@ -105849,18 +105849,18 @@ var require_zip_archive_output_stream = __commonJS({ ZipArchiveOutputStream.prototype._smartStream = function(ae, callback) { var deflate = ae.getMethod() === constants.METHOD_DEFLATED; var process2 = deflate ? new DeflateCRC32Stream(this.options.zlib) : new CRC32Stream(); - var error4 = null; + var error3 = null; function handleStuff() { var digest = process2.digest().readUInt32BE(0); ae.setCrc(digest); ae.setSize(process2.size()); ae.setCompressedSize(process2.size(true)); this._afterAppend(ae); - callback(error4, ae); + callback(error3, ae); } process2.once("end", handleStuff.bind(this)); process2.once("error", function(err) { - error4 = err; + error3 = err; }); process2.pipe(this, { end: false }); return process2; @@ -107226,11 +107226,11 @@ var require_streamx = __commonJS({ } [asyncIterator]() { const stream2 = this; - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("readable", onreadable); this.on("close", onclose); @@ -107262,7 +107262,7 @@ var require_streamx = __commonJS({ } function ondata(data) { if (promiseReject === null) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else if (data === null && (stream2._duplexState & READ_DONE) === 0) promiseReject(STREAM_DESTROYED); else promiseResolve({ value: data, done: data === null }); promiseReject = promiseResolve = null; @@ -107436,7 +107436,7 @@ var require_streamx = __commonJS({ if (all.length < 2) throw new Error("Pipeline requires at least 2 streams"); let src = all[0]; let dest = null; - let error4 = null; + let error3 = null; for (let i = 1; i < all.length; i++) { dest = all[i]; if (isStreamx(src)) { @@ -107451,14 +107451,14 @@ var require_streamx = __commonJS({ let fin = false; const autoDestroy = isStreamx(dest) || !!(dest._writableState && dest._writableState.autoDestroy); dest.on("error", (err) => { - if (error4 === null) error4 = err; + if (error3 === null) error3 = err; }); dest.on("finish", () => { fin = true; - if (!autoDestroy) done(error4); + if (!autoDestroy) done(error3); }); if (autoDestroy) { - dest.on("close", () => done(error4 || (fin ? null : PREMATURE_CLOSE))); + dest.on("close", () => done(error3 || (fin ? null : PREMATURE_CLOSE))); } } return dest; @@ -107471,8 +107471,8 @@ var require_streamx = __commonJS({ } } function onerror(err) { - if (!err || error4) return; - error4 = err; + if (!err || error3) return; + error3 = err; for (const s of all) { s.destroy(err); } @@ -108038,7 +108038,7 @@ var require_extract = __commonJS({ cb(null); } [Symbol.asyncIterator]() { - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; let entryStream = null; @@ -108046,7 +108046,7 @@ var require_extract = __commonJS({ const extract2 = this; this.on("entry", onentry); this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("close", onclose); return { @@ -108070,8 +108070,8 @@ var require_extract = __commonJS({ cb(err); } function onnext(resolve8, reject) { - if (error4) { - return reject(error4); + if (error3) { + return reject(error3); } if (entryStream) { resolve8({ value: entryStream, done: false }); @@ -108097,9 +108097,9 @@ var require_extract = __commonJS({ } } function onclose() { - consumeCallback(error4); + consumeCallback(error3); if (!promiseResolve) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else promiseResolve({ value: void 0, done: true }); promiseResolve = promiseReject = null; } @@ -108995,18 +108995,18 @@ var require_zip2 = __commonJS({ return zipUploadStream; }); } - var zipErrorCallback = (error4) => { + var zipErrorCallback = (error3) => { core18.error("An error has occurred while creating the zip file for upload"); - core18.info(error4); + core18.info(error3); throw new Error("An error has occurred during zip creation for the artifact"); }; - var zipWarningCallback = (error4) => { - if (error4.code === "ENOENT") { + var zipWarningCallback = (error3) => { + if (error3.code === "ENOENT") { core18.warning("ENOENT warning during artifact zip creation. No such file or directory"); - core18.info(error4); + core18.info(error3); } else { - core18.warning(`A non-blocking warning has occurred during artifact zip creation: ${error4.code}`); - core18.info(error4); + core18.warning(`A non-blocking warning has occurred during artifact zip creation: ${error3.code}`); + core18.info(error3); } }; var zipFinishCallback = () => { @@ -110813,8 +110813,8 @@ var require_parser_stream = __commonJS({ this.unzipStream.on("entry", function(entry) { self2.push(entry); }); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(ParserStream, Transform); @@ -110954,8 +110954,8 @@ var require_extract2 = __commonJS({ this.createdDirectories = {}; var self2 = this; this.unzipStream.on("entry", this._processEntry.bind(this)); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(Extract, Transform); @@ -110989,8 +110989,8 @@ var require_extract2 = __commonJS({ self2.unfinishedEntries--; self2._notifyAwaiter(); }); - pipedStream.on("error", function(error4) { - self2.emit("error", error4); + pipedStream.on("error", function(error3) { + self2.emit("error", error3); }); entry.pipe(pipedStream); }; @@ -111125,11 +111125,11 @@ var require_download_artifact = __commonJS({ try { yield promises_1.default.access(path15); return true; - } catch (error4) { - if (error4.code === "ENOENT") { + } catch (error3) { + if (error3.code === "ENOENT") { return false; } else { - throw error4; + throw error3; } } }); @@ -111140,9 +111140,9 @@ var require_download_artifact = __commonJS({ while (retryCount < 5) { try { return yield streamExtractExternal(url2, directory); - } catch (error4) { + } catch (error3) { retryCount++; - core18.debug(`Failed to download artifact after ${retryCount} retries due to ${error4.message}. Retrying in 5 seconds...`); + core18.debug(`Failed to download artifact after ${retryCount} retries due to ${error3.message}. Retrying in 5 seconds...`); yield new Promise((resolve8) => setTimeout(resolve8, 5e3)); } } @@ -111171,10 +111171,10 @@ var require_download_artifact = __commonJS({ const extractStream = passThrough; extractStream.on("data", () => { timer.refresh(); - }).on("error", (error4) => { - core18.debug(`response.message: Artifact download failed: ${error4.message}`); + }).on("error", (error3) => { + core18.debug(`response.message: Artifact download failed: ${error3.message}`); clearTimeout(timer); - reject(error4); + reject(error3); }).pipe(unzip_stream_1.default.Extract({ path: directory })).on("close", () => { clearTimeout(timer); if (hashStream) { @@ -111183,8 +111183,8 @@ var require_download_artifact = __commonJS({ core18.info(`SHA256 digest of downloaded artifact is ${sha256Digest}`); } resolve8({ sha256Digest: `sha256:${sha256Digest}` }); - }).on("error", (error4) => { - reject(error4); + }).on("error", (error3) => { + reject(error3); }); }); }); @@ -111223,8 +111223,8 @@ var require_download_artifact = __commonJS({ core18.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -111266,8 +111266,8 @@ Are you trying to download from a different run? Try specifying a github-token w core18.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -111365,9 +111365,9 @@ var require_dist_node16 = __commonJS({ return request(options).then((response) => { octokit.log.info(`${requestOptions.method} ${path15} - ${response.status} in ${Date.now() - start}ms`); return response; - }).catch((error4) => { - octokit.log.info(`${requestOptions.method} ${path15} - ${error4.status} in ${Date.now() - start}ms`); - throw error4; + }).catch((error3) => { + octokit.log.info(`${requestOptions.method} ${path15} - ${error3.status} in ${Date.now() - start}ms`); + throw error3; }); }); } @@ -111385,22 +111385,22 @@ var require_dist_node17 = __commonJS({ return ex && typeof ex === "object" && "default" in ex ? ex["default"] : ex; } var Bottleneck = _interopDefault(require_light()); - async function errorRequest(octokit, state, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(octokit, state, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } async function wrapRequest(state, request, options) { const limiter = new Bottleneck(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -111422,12 +111422,12 @@ var require_dist_node17 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -111918,13 +111918,13 @@ var require_client2 = __commonJS({ throw new errors_1.GHESNotSupportedError(); } return (0, upload_artifact_1.uploadArtifact)(name, files, rootDirectory, options); - } catch (error4) { - (0, core_1.warning)(`Artifact upload failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Artifact upload failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions is operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -111939,13 +111939,13 @@ If the error persists, please check whether Actions is operating normally at [ht return (0, download_artifact_1.downloadArtifactPublic)(artifactId, repositoryOwner, repositoryName, token, downloadOptions); } return (0, download_artifact_1.downloadArtifactInternal)(artifactId, options); - } catch (error4) { - (0, core_1.warning)(`Download Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Download Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -111960,13 +111960,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, list_artifacts_1.listArtifactsPublic)(workflowRunId, repositoryOwner, repositoryName, token, options === null || options === void 0 ? void 0 : options.latest); } return (0, list_artifacts_1.listArtifactsInternal)(options === null || options === void 0 ? void 0 : options.latest); - } catch (error4) { - (0, core_1.warning)(`Listing Artifacts failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Listing Artifacts failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -111981,13 +111981,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, get_artifact_1.getArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, get_artifact_1.getArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Get Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Get Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -112002,13 +112002,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, delete_artifact_1.deleteArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, delete_artifact_1.deleteArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Delete Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Delete Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -112508,14 +112508,14 @@ var require_tmp = __commonJS({ options.template = _getRelativePathSync("template", options.template, tmpDir); return options; } - function _isEBADF(error4) { - return _isExpectedError(error4, -EBADF, "EBADF"); + function _isEBADF(error3) { + return _isExpectedError(error3, -EBADF, "EBADF"); } - function _isENOENT(error4) { - return _isExpectedError(error4, -ENOENT, "ENOENT"); + function _isENOENT(error3) { + return _isExpectedError(error3, -ENOENT, "ENOENT"); } - function _isExpectedError(error4, errno, code) { - return IS_WIN32 ? error4.code === code : error4.code === code && error4.errno === errno; + function _isExpectedError(error3, errno, code) { + return IS_WIN32 ? error3.code === code : error3.code === code && error3.errno === errno; } function setGracefulCleanup() { _gracefulCleanup = true; @@ -114227,9 +114227,9 @@ var require_upload_gzip = __commonJS({ const size = (yield stat(tempFilePath)).size; resolve8(size); })); - outputStream.on("error", (error4) => { - console.log(error4); - reject(error4); + outputStream.on("error", (error3) => { + console.log(error3); + reject(error3); }); }); }); @@ -114354,9 +114354,9 @@ var require_requestUtils2 = __commonJS({ } isRetryable = (0, utils_1.isRetryableStatusCode)(statusCode); errorMessage = `Artifact service responded with ${statusCode}`; - } catch (error4) { + } catch (error3) { isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { core18.info(`${name} - Error is not retryable`); @@ -114722,9 +114722,9 @@ var require_upload_http_client = __commonJS({ let response; try { response = yield uploadChunkRequest(); - } catch (error4) { + } catch (error3) { core18.info(`An error has been caught http-client index ${httpClientIndex}, retrying the upload`); - console.log(error4); + console.log(error3); if (incrementAndCheckRetryLimit()) { return false; } @@ -114913,8 +114913,8 @@ var require_download_http_client = __commonJS({ } this.statusReporter.incrementProcessedCount(); } - }))).catch((error4) => { - throw new Error(`Unable to download the artifact: ${error4}`); + }))).catch((error3) => { + throw new Error(`Unable to download the artifact: ${error3}`); }).finally(() => { this.statusReporter.stop(); this.downloadHttpManager.disposeAndReplaceAllClients(); @@ -114979,9 +114979,9 @@ var require_download_http_client = __commonJS({ let response; try { response = yield makeDownloadRequest(); - } catch (error4) { + } catch (error3) { core18.info("An error occurred while attempting to download a file"); - console.log(error4); + console.log(error3); yield backOff(); continue; } @@ -114995,7 +114995,7 @@ var require_download_http_client = __commonJS({ } else { forceRetry = true; } - } catch (error4) { + } catch (error3) { forceRetry = true; } } @@ -115021,31 +115021,31 @@ var require_download_http_client = __commonJS({ yield new Promise((resolve8, reject) => { if (isGzip) { const gunzip = zlib3.createGunzip(); - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core18.info(`An error occurred while attempting to read the response stream`); gunzip.close(); destinationStream.close(); - reject(error4); - }).pipe(gunzip).on("error", (error4) => { + reject(error3); + }).pipe(gunzip).on("error", (error3) => { core18.info(`An error occurred while attempting to decompress the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve8(); - }).on("error", (error4) => { + }).on("error", (error3) => { core18.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } else { - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core18.info(`An error occurred while attempting to read the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve8(); - }).on("error", (error4) => { + }).on("error", (error3) => { core18.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } }); @@ -119362,14 +119362,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs17.lstat(itemPath, { bigint: true }) : await fs17.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs17.lstat(itemPath, { bigint: true }) : await fs17.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs17.readdir(itemPath) : await fs17.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs17.readdir(itemPath) : await fs17.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -119380,13 +119380,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -122008,9 +122008,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -122204,11 +122204,11 @@ function parseMatrixInput(matrixInput) { } return JSON.parse(matrixInput); } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } async function checkDiskUsage(logger) { try { @@ -122231,9 +122231,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -122717,19 +122717,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -122745,9 +122745,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -123007,13 +123007,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -125644,9 +125644,9 @@ var JobStatus = /* @__PURE__ */ ((JobStatus2) => { JobStatus2["ConfigErrorStatus"] = "JOB_STATUS_CONFIGURATION_ERROR"; return JobStatus2; })(JobStatus || {}); -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -127279,15 +127279,15 @@ function validateSarifFileSchema(sarif, sarifFilePath, logger) { const warnings = (result.errors ?? []).filter( (err) => err.name === "format" && typeof err.argument === "string" && warningAttributes.includes(err.argument) ); - for (const warning11 of warnings) { + for (const warning12 of warnings) { logger.info( - `Warning: '${warning11.instance}' is not a valid URI in '${warning11.property}'.` + `Warning: '${warning12.instance}' is not a valid URI in '${warning12.property}'.` ); } if (errors.length > 0) { - for (const error4 of errors) { - logger.startGroup(`Error details: ${error4.stack}`); - logger.info(JSON.stringify(error4, null, 2)); + for (const error3 of errors) { + logger.startGroup(`Error details: ${error3.stack}`); + logger.info(JSON.stringify(error3, null, 2)); logger.endGroup(); } const sarifErrors = errors.map((e) => `- ${e.stack}`); @@ -127512,10 +127512,10 @@ function shouldConsiderConfigurationError(processingErrors) { } function shouldConsiderInvalidRequest(processingErrors) { return processingErrors.every( - (error4) => error4.startsWith("rejecting SARIF") || error4.startsWith("an invalid URI was provided as a SARIF location") || error4.startsWith("locationFromSarifResult: expected artifact location") || error4.startsWith( + (error3) => error3.startsWith("rejecting SARIF") || error3.startsWith("an invalid URI was provided as a SARIF location") || error3.startsWith("locationFromSarifResult: expected artifact location") || error3.startsWith( "could not convert rules: invalid security severity value, is not a number" ) || /^SARIF URI scheme [^\s]* did not match the checkout URI scheme [^\s]*/.test( - error4 + error3 ) ); } @@ -127726,8 +127726,8 @@ function getCheckoutPathInputOrThrow(workflow, jobName, matrixVars) { } // src/init-action-post-helper.ts -function createFailedUploadFailedSarifResult(error4) { - const wrappedError = wrapError(error4); +function createFailedUploadFailedSarifResult(error3) { + const wrappedError = wrapError(error3); return { upload_failed_run_error: wrappedError.message, upload_failed_run_stack_trace: wrappedError.stack @@ -127820,9 +127820,9 @@ async function run(uploadAllAvailableDebugArtifacts, printDebugLogs2, codeql, co ); } if (process.env["CODEQL_ACTION_EXPECT_UPLOAD_FAILED_SARIF"] === "true" && !uploadFailedSarifResult.raw_upload_size_bytes) { - const error4 = JSON.stringify(uploadFailedSarifResult); + const error3 = JSON.stringify(uploadFailedSarifResult); throw new Error( - `Expected to upload a failed SARIF file for this CodeQL code scanning run, but the result was instead ${error4}.` + `Expected to upload a failed SARIF file for this CodeQL code scanning run, but the result was instead ${error3}.` ); } if (process.env["CODEQL_ACTION_EXPECT_UPLOAD_FAILED_SARIF"] === "true") { @@ -127975,17 +127975,17 @@ async function runWrapper() { } } } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); - core17.setFailed(error4.message); + const error3 = wrapError(unwrappedError); + core17.setFailed(error3.message); const statusReportBase2 = await createStatusReportBase( "init-post" /* InitPost */, - getActionsStatus(error4), + getActionsStatus(error3), startedAt, config, await checkDiskUsage(logger), logger, - error4.message, - error4.stack + error3.message, + error3.stack ); if (statusReportBase2 !== void 0) { await sendStatusReport(statusReportBase2); diff --git a/lib/init-action.js b/lib/init-action.js index 2c35e5c68..6de0e4e2d 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug5("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve9(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug3() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug5; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning10(message, properties = {}) { + exports2.error = error3; + function warning11(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning10; + exports2.warning = warning11; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -23133,8 +23133,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -23866,7 +23866,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -23875,7 +23875,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -23885,17 +23885,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24514,7 +24514,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -24523,7 +24523,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -24533,17 +24533,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -27215,9 +27215,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -27911,8 +27911,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27922,8 @@ var require_light = __commonJS({ return (await Promise.all(promises3)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28035,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28072,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run2, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28090,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } } doExpire(clearGlobalState, run2, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run2, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run2, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28117,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28396,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve9, returned, task; + var args, cb, error3, reject, resolve9, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve9, reject } = this._queue.shift()); @@ -28407,9 +28407,9 @@ var require_light = __commonJS({ return resolve9(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28543,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28877,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29180,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29213,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29238,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -36236,8 +36236,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36471,9 +36471,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37804,14 +37804,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37821,7 +37821,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37829,8 +37829,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -38050,7 +38050,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39260,14 +39260,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39939,11 +39939,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39958,8 +39958,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40453,8 +40453,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40688,9 +40688,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41190,8 +41190,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41425,9 +41425,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42490,9 +42490,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42540,13 +42540,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42566,21 +42566,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42745,8 +42745,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43152,16 +43152,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43705,10 +43705,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45787,12 +45787,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -46057,16 +46057,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46191,7 +46191,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46357,7 +46357,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46598,9 +46598,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47308,7 +47308,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48371,26 +48371,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48431,12 +48431,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48444,13 +48444,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48459,7 +48459,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -63065,8 +63065,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -63100,10 +63100,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64687,8 +64687,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64703,9 +64703,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve9, reject) => { this.emitter.on("finish", resolve9); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65806,8 +65806,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65891,7 +65891,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -69043,7 +69043,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70410,9 +70410,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70526,12 +70526,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70565,13 +70565,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71387,8 +71387,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73234,8 +73234,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74806,12 +74806,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74831,12 +74831,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75300,8 +75300,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75312,8 +75312,8 @@ var require_test_transport = __commonJS({ stream2.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } } @@ -76376,8 +76376,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76473,8 +76473,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76505,18 +76505,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76784,8 +76784,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76986,22 +76986,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -77056,15 +77056,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -77072,8 +77072,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -77135,10 +77135,10 @@ var require_cache3 = __commonJS({ } core14.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else { @@ -77151,8 +77151,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77197,8 +77197,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core14.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core14.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError2(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core14.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77217,10 +77217,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77235,8 +77235,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -81059,7 +81059,7 @@ var require_debug2 = __commonJS({ if (!debug5) { try { debug5 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug5 !== "function") { debug5 = function() { @@ -81092,8 +81092,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -81167,9 +81167,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -81336,10 +81336,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -81538,12 +81538,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -81660,14 +81660,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs15.lstat(itemPath, { bigint: true }) : await fs15.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs15.lstat(itemPath, { bigint: true }) : await fs15.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs15.readdir(itemPath) : await fs15.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs15.readdir(itemPath) : await fs15.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -81678,13 +81678,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -84310,9 +84310,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -84700,11 +84700,11 @@ function parseMatrixInput(matrixInput) { } return JSON.parse(matrixInput); } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } function prettyPrintPack(pack) { return `${pack.name}${pack.version ? `@${pack.version}` : ""}${pack.path ? `:${pack.path}` : ""}`; @@ -84730,9 +84730,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -84744,7 +84744,7 @@ function checkActionVersion(version, githubVersion) { semver.coerce(githubVersion.version) ?? "0.0.0", ">=3.20" )) { - core3.error( + core3.warning( "CodeQL Action v3 will be deprecated in December 2026. Please update all occurrences of the CodeQL Action in your workflow files to v4. For more information, see https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/" ); core3.exportVariable("CODEQL_ACTION_DID_LOG_VERSION_DEPRECATION" /* LOG_VERSION_DEPRECATION */, "true"); @@ -85309,9 +85309,9 @@ function getInvalidConfigFileMessage(configFile, messages) { return `The configuration file "${configFile}" is invalid: ${messages.slice(0, 10).join(", ")}${andMore}`; } function getConfigFileRepoFormatInvalidMessage(configFile) { - let error4 = `The configuration file "${configFile}" is not a supported remote file reference.`; - error4 += " Expected format //@"; - return error4; + let error3 = `The configuration file "${configFile}" is not a supported remote file reference.`; + error3 += " Expected format //@"; + return error3; } function getConfigFileFormatInvalidMessage(configFile) { return `The configuration file "${configFile}" could not be read`; @@ -85322,15 +85322,15 @@ function getConfigFileDirectoryGivenMessage(configFile) { function getEmptyCombinesError() { return `A '+' was used to specify that you want to add extra arguments to the configuration, but no extra arguments were specified. Please either remove the '+' or specify some extra arguments.`; } -function getConfigFilePropertyError(configFile, property, error4) { +function getConfigFilePropertyError(configFile, property, error3) { if (configFile === void 0) { - return `The workflow property "${property}" is invalid: ${error4}`; + return `The workflow property "${property}" is invalid: ${error3}`; } else { - return `The configuration file "${configFile}" is invalid: property "${property}" ${error4}`; + return `The configuration file "${configFile}" is invalid: property "${property}" ${error3}`; } } -function getRepoPropertyError(propertyName, error4) { - return `The repository property "${propertyName}" is invalid: ${error4}`; +function getRepoPropertyError(propertyName, error3) { + return `The repository property "${propertyName}" is invalid: ${error3}`; } function getPacksStrInvalid(packStr, configFile) { return configFile ? getConfigFilePropertyError( @@ -85609,8 +85609,8 @@ function parseUserConfig(logger, pathInput, contents, validateConfig) { if (validateConfig) { const result = new jsonschema.Validator().validate(doc, schema2); if (result.errors.length > 0) { - for (const error4 of result.errors) { - logger.error(error4.stack); + for (const error3 of result.errors) { + logger.error(error3.stack); } throw new ConfigurationError( getInvalidConfigFileMessage( @@ -85621,13 +85621,13 @@ function parseUserConfig(logger, pathInput, contents, validateConfig) { } } return doc; - } catch (error4) { - if (error4 instanceof YAMLException) { + } catch (error3) { + if (error3 instanceof YAMLException) { throw new ConfigurationError( - getConfigFileParseErrorMessage(pathInput, error4.message) + getConfigFileParseErrorMessage(pathInput, error3.message) ); } - throw error4; + throw error3; } } @@ -85667,13 +85667,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -86005,9 +86005,9 @@ async function downloadOverlayBaseDatabaseFromCache(codeql, config, logger) { logger.info( `Downloaded overlay-base database in cache with key ${foundKey}` ); - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to download overlay-base database from cache: ${error4 instanceof Error ? error4.message : String(error4)}` + `Failed to download overlay-base database from cache: ${error3 instanceof Error ? error3.message : String(error3)}` ); return void 0; } @@ -87521,19 +87521,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -87549,9 +87549,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -89382,9 +89382,9 @@ function isFirstPartyAnalysis(actionName) { } return process.env["CODEQL_ACTION_INIT_HAS_RUN" /* INIT_ACTION_HAS_RUN */] === "true"; } -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -89806,16 +89806,16 @@ async function sendStartingStatusReport(startedAt, config, logger) { await sendStatusReport(statusReportBase); } } -async function sendCompletedStatusReport(startedAt, config, configFile, toolsDownloadStatusReport, toolsFeatureFlagsValid, toolsSource, toolsVersion, overlayBaseDatabaseStats, dependencyCachingResults, logger, error4) { +async function sendCompletedStatusReport(startedAt, config, configFile, toolsDownloadStatusReport, toolsFeatureFlagsValid, toolsSource, toolsVersion, overlayBaseDatabaseStats, dependencyCachingResults, logger, error3) { const statusReportBase = await createStatusReportBase( "init" /* Init */, - getActionsStatus(error4), + getActionsStatus(error3), startedAt, config, await checkDiskUsage(logger), logger, - error4?.message, - error4?.stack + error3?.message, + error3?.stack ); if (statusReportBase === void 0) { return; @@ -89979,17 +89979,17 @@ async function run() { }); await checkInstallPython311(config.languages, codeql); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); - core13.setFailed(error4.message); + const error3 = wrapError(unwrappedError); + core13.setFailed(error3.message); const statusReportBase = await createStatusReportBase( "init" /* Init */, - error4 instanceof ConfigurationError ? "user-error" : "aborted", + error3 instanceof ConfigurationError ? "user-error" : "aborted", startedAt, config, await checkDiskUsage(logger), logger, - error4.message, - error4.stack + error3.message, + error3.stack ); if (statusReportBase !== void 0) { await sendStatusReport(statusReportBase); @@ -90232,8 +90232,8 @@ exec ${goBinaryPath} "$@"` core13.setOutput("codeql-path", config.codeQLCmd); core13.setOutput("codeql-version", (await codeql.getVersion()).version); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); - core13.setFailed(error4.message); + const error3 = wrapError(unwrappedError); + core13.setFailed(error3.message); await sendCompletedStatusReport( startedAt, config, @@ -90246,7 +90246,7 @@ exec ${goBinaryPath} "$@"` overlayBaseDatabaseStats, dependencyCachingResults, logger, - error4 + error3 ); return; } finally { @@ -90295,8 +90295,8 @@ async function recordZstdAvailability(config, zstdAvailability) { async function runWrapper() { try { await run(); - } catch (error4) { - core13.setFailed(`init action failed: ${getErrorMessage(error4)}`); + } catch (error3) { + core13.setFailed(`init action failed: ${getErrorMessage(error3)}`); } await checkForTimeout(); } diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index 4059c9db1..8a757d8c6 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug5("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve4(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug5; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning8(message, properties = {}) { + exports2.error = error3; + function warning9(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning8; + exports2.warning = warning9; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -27911,8 +27911,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27922,8 @@ var require_light = __commonJS({ return (await Promise.all(promises2)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28035,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28072,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run2, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28090,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } } doExpire(clearGlobalState, run2, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run2, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run2, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28117,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28396,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve4, returned, task; + var args, cb, error3, reject, resolve4, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve4, reject } = this._queue.shift()); @@ -28407,9 +28407,9 @@ var require_light = __commonJS({ return resolve4(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28543,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28877,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29180,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29213,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29238,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -36085,8 +36085,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36320,9 +36320,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37653,14 +37653,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37670,7 +37670,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37678,8 +37678,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -37899,7 +37899,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39109,14 +39109,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39788,11 +39788,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39807,8 +39807,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40302,8 +40302,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40537,9 +40537,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41039,8 +41039,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41274,9 +41274,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42339,9 +42339,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42389,13 +42389,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42415,21 +42415,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42594,8 +42594,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43001,16 +43001,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43554,10 +43554,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45636,12 +45636,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -45906,16 +45906,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46040,7 +46040,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46206,7 +46206,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46447,9 +46447,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47157,7 +47157,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48220,26 +48220,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48280,12 +48280,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48293,13 +48293,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48308,7 +48308,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -62914,8 +62914,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -62949,10 +62949,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64536,8 +64536,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64552,9 +64552,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve4, reject) => { this.emitter.on("finish", resolve4); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65655,8 +65655,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65740,7 +65740,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -68892,7 +68892,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70259,9 +70259,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core13.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core13.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70375,12 +70375,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70414,13 +70414,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71236,8 +71236,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73083,8 +73083,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74655,12 +74655,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74680,12 +74680,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75149,8 +75149,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75161,8 +75161,8 @@ var require_test_transport = __commonJS({ stream.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } } @@ -76225,8 +76225,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76322,8 +76322,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76354,18 +76354,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76633,8 +76633,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76835,22 +76835,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core13.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core13.error(`Failed to restore: ${error4.message}`); + core13.error(`Failed to restore: ${error3.message}`); } else { - core13.warning(`Failed to restore: ${error4.message}`); + core13.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76905,15 +76905,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core13.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core13.error(`Failed to restore: ${error4.message}`); + core13.error(`Failed to restore: ${error3.message}`); } else { - core13.warning(`Failed to restore: ${error4.message}`); + core13.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -76921,8 +76921,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76984,10 +76984,10 @@ var require_cache3 = __commonJS({ } core13.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core13.info(`Failed to save: ${typedError.message}`); } else { @@ -77000,8 +77000,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77046,8 +77046,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core13.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core13.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core13.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77066,10 +77066,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core13.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77084,8 +77084,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79811,7 @@ var require_debug2 = __commonJS({ if (!debug5) { try { debug5 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug5 !== "function") { debug5 = function() { @@ -79844,8 +79844,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79919,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80088,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80290,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -80351,14 +80351,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs5.lstat(itemPath, { bigint: true }) : await fs5.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs5.lstat(itemPath, { bigint: true }) : await fs5.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs5.readdir(itemPath) : await fs5.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs5.readdir(itemPath) : await fs5.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -80369,13 +80369,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -82996,9 +82996,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -83138,11 +83138,11 @@ async function checkForTimeout() { process.exit(); } } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } async function checkDiskUsage(logger) { try { @@ -83165,9 +83165,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -83179,7 +83179,7 @@ function checkActionVersion(version, githubVersion) { semver.coerce(githubVersion.version) ?? "0.0.0", ">=3.20" )) { - core3.error( + core3.warning( "CodeQL Action v3 will be deprecated in December 2026. Please update all occurrences of the CodeQL Action in your workflow files to v4. For more information, see https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/" ); core3.exportVariable("CODEQL_ACTION_DID_LOG_VERSION_DEPRECATION" /* LOG_VERSION_DEPRECATION */, "true"); @@ -83461,19 +83461,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -83489,9 +83489,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -83726,13 +83726,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -84812,9 +84812,9 @@ function isFirstPartyAnalysis(actionName) { } return process.env["CODEQL_ACTION_INIT_HAS_RUN" /* INIT_ACTION_HAS_RUN */] === "true"; } -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -85020,25 +85020,25 @@ async function run() { ); core12.setOutput(ENVIRONMENT_OUTPUT_NAME, result); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); - if (error4 instanceof CliError) { + const error3 = wrapError(unwrappedError); + if (error3 instanceof CliError) { core12.setOutput(ENVIRONMENT_OUTPUT_NAME, {}); logger.warning( - `Failed to resolve a build environment suitable for automatically building your code. ${error4.message}` + `Failed to resolve a build environment suitable for automatically building your code. ${error3.message}` ); } else { core12.setFailed( - `Failed to resolve a build environment suitable for automatically building your code. ${error4.message}` + `Failed to resolve a build environment suitable for automatically building your code. ${error3.message}` ); const statusReportBase2 = await createStatusReportBase( "resolve-environment" /* ResolveEnvironment */, - getActionsStatus(error4), + getActionsStatus(error3), startedAt, config, await checkDiskUsage(logger), logger, - error4.message, - error4.stack + error3.message, + error3.stack ); if (statusReportBase2 !== void 0) { await sendStatusReport(statusReportBase2); @@ -85061,10 +85061,10 @@ async function run() { async function runWrapper() { try { await run(); - } catch (error4) { + } catch (error3) { core12.setFailed( `${"resolve-environment" /* ResolveEnvironment */} action failed: ${getErrorMessage( - error4 + error3 )}` ); } diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index 3caba058d..31f135f53 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug5("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve4(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug5; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning8(message, properties = {}) { + exports2.error = error3; + function warning9(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning8; + exports2.warning = warning9; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -27911,8 +27911,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27922,8 @@ var require_light = __commonJS({ return (await Promise.all(promises3)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28035,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28072,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run2, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28090,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } } doExpire(clearGlobalState, run2, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run2, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run2, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28117,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28396,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve4, returned, task; + var args, cb, error3, reject, resolve4, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve4, reject } = this._queue.shift()); @@ -28407,9 +28407,9 @@ var require_light = __commonJS({ return resolve4(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28543,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28877,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29180,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29213,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29238,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -34788,8 +34788,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -35023,9 +35023,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -36356,14 +36356,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -36373,7 +36373,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -36381,8 +36381,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -36602,7 +36602,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -37812,14 +37812,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -38491,11 +38491,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -38510,8 +38510,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -39005,8 +39005,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -39240,9 +39240,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -39742,8 +39742,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -39977,9 +39977,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -41042,9 +41042,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -41092,13 +41092,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -41118,21 +41118,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -41297,8 +41297,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -41704,16 +41704,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -42257,10 +42257,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -44339,12 +44339,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -44609,16 +44609,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -44743,7 +44743,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -44909,7 +44909,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -45150,9 +45150,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -45860,7 +45860,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -46923,26 +46923,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -46983,12 +46983,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -46996,13 +46996,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -47011,7 +47011,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -61617,8 +61617,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -61652,10 +61652,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -63239,8 +63239,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -63255,9 +63255,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve4, reject) => { this.emitter.on("finish", resolve4); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -64358,8 +64358,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -64443,7 +64443,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -67595,7 +67595,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -68962,9 +68962,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core13.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core13.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -69078,12 +69078,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -69117,13 +69117,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -69939,8 +69939,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -71786,8 +71786,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -73358,12 +73358,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -73383,12 +73383,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -73852,8 +73852,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -73864,8 +73864,8 @@ var require_test_transport = __commonJS({ stream2.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } } @@ -74928,8 +74928,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -75025,8 +75025,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -75057,18 +75057,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -75336,8 +75336,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -75538,22 +75538,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core13.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core13.error(`Failed to restore: ${error4.message}`); + core13.error(`Failed to restore: ${error3.message}`); } else { - core13.warning(`Failed to restore: ${error4.message}`); + core13.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -75608,15 +75608,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core13.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core13.error(`Failed to restore: ${error4.message}`); + core13.error(`Failed to restore: ${error3.message}`); } else { - core13.warning(`Failed to restore: ${error4.message}`); + core13.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -75624,8 +75624,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -75687,10 +75687,10 @@ var require_cache3 = __commonJS({ } core13.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core13.info(`Failed to save: ${typedError.message}`); } else { @@ -75703,8 +75703,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -75749,8 +75749,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core13.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core13.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core13.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -75769,10 +75769,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core13.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -75787,8 +75787,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79811,7 @@ var require_debug2 = __commonJS({ if (!debug5) { try { debug5 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug5 !== "function") { debug5 = function() { @@ -79844,8 +79844,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79919,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80088,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80290,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -80407,14 +80407,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs9.lstat(itemPath, { bigint: true }) : await fs9.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs9.lstat(itemPath, { bigint: true }) : await fs9.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs9.readdir(itemPath) : await fs9.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs9.readdir(itemPath) : await fs9.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -80425,13 +80425,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -83053,9 +83053,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -83214,11 +83214,11 @@ async function checkForTimeout() { process.exit(); } } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } async function checkDiskUsage(logger) { try { @@ -83241,9 +83241,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -83255,7 +83255,7 @@ function checkActionVersion(version, githubVersion) { semver.coerce(githubVersion.version) ?? "0.0.0", ">=3.20" )) { - core3.error( + core3.warning( "CodeQL Action v3 will be deprecated in December 2026. Please update all occurrences of the CodeQL Action in your workflow files to v4. For more information, see https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/" ); core3.exportVariable("CODEQL_ACTION_DID_LOG_VERSION_DEPRECATION" /* LOG_VERSION_DEPRECATION */, "true"); @@ -83624,13 +83624,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -84355,19 +84355,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -84383,9 +84383,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -86133,9 +86133,9 @@ function isFirstPartyAnalysis(actionName) { } return process.env["CODEQL_ACTION_INIT_HAS_RUN" /* INIT_ACTION_HAS_RUN */] === "true"; } -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -86306,16 +86306,16 @@ async function sendStatusReport(statusReport) { } // src/setup-codeql-action.ts -async function sendCompletedStatusReport(startedAt, toolsDownloadStatusReport, toolsFeatureFlagsValid, toolsSource, toolsVersion, logger, error4) { +async function sendCompletedStatusReport(startedAt, toolsDownloadStatusReport, toolsFeatureFlagsValid, toolsSource, toolsVersion, logger, error3) { const statusReportBase = await createStatusReportBase( "setup-codeql" /* SetupCodeQL */, - getActionsStatus(error4), + getActionsStatus(error3), startedAt, void 0, await checkDiskUsage(logger), logger, - error4?.message, - error4?.stack + error3?.message, + error3?.stack ); if (statusReportBase === void 0) { return; @@ -86397,17 +86397,17 @@ async function run() { core12.setOutput("codeql-version", (await codeql.getVersion()).version); core12.exportVariable("CODEQL_ACTION_SETUP_CODEQL_HAS_RUN" /* SETUP_CODEQL_ACTION_HAS_RUN */, "true"); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); - core12.setFailed(error4.message); + const error3 = wrapError(unwrappedError); + core12.setFailed(error3.message); const statusReportBase = await createStatusReportBase( "setup-codeql" /* SetupCodeQL */, - error4 instanceof ConfigurationError ? "user-error" : "failure", + error3 instanceof ConfigurationError ? "user-error" : "failure", startedAt, void 0, await checkDiskUsage(logger), logger, - error4.message, - error4.stack + error3.message, + error3.stack ); if (statusReportBase !== void 0) { await sendStatusReport(statusReportBase); @@ -86426,8 +86426,8 @@ async function run() { async function runWrapper() { try { await run(); - } catch (error4) { - core12.setFailed(`setup-codeql action failed: ${getErrorMessage(error4)}`); + } catch (error3) { + core12.setFailed(`setup-codeql action failed: ${getErrorMessage(error3)}`); } await checkForTimeout(); } diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index 3c7c44495..86991078a 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug4("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay === "number" && delay > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve2(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed; function isDebug3() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug4; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning8(message, properties = {}) { + exports2.error = error3; + function warning9(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning8; + exports2.warning = warning9; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -27911,8 +27911,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27922,8 @@ var require_light = __commonJS({ return (await Promise.all(promises)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28035,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28072,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28090,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } } doExpire(clearGlobalState, run, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28117,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28396,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve2, returned, task; + var args, cb, error3, reject, resolve2, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve2, reject } = this._queue.shift()); @@ -28407,9 +28407,9 @@ var require_light = __commonJS({ return resolve2(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28543,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28877,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29180,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29213,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29238,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -36085,8 +36085,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36320,9 +36320,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37653,14 +37653,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37670,7 +37670,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37678,8 +37678,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -37899,7 +37899,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39109,14 +39109,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39788,11 +39788,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39807,8 +39807,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40302,8 +40302,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40537,9 +40537,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41039,8 +41039,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41274,9 +41274,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42339,9 +42339,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42389,13 +42389,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42415,21 +42415,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42594,8 +42594,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43001,16 +43001,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43554,10 +43554,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45636,12 +45636,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -45906,16 +45906,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46040,7 +46040,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46206,7 +46206,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46447,9 +46447,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47157,7 +47157,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48220,26 +48220,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48280,12 +48280,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48293,13 +48293,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48308,7 +48308,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -62914,8 +62914,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -62949,10 +62949,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64536,8 +64536,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64552,9 +64552,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve2, reject) => { this.emitter.on("finish", resolve2); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65655,8 +65655,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65740,7 +65740,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -68892,7 +68892,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70259,9 +70259,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70375,12 +70375,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70414,13 +70414,13 @@ var require_requestUtils = __commonJS({ delay, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71236,8 +71236,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73083,8 +73083,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74655,12 +74655,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74680,12 +74680,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75149,8 +75149,8 @@ var require_test_transport = __commonJS({ } try { yield delay(this.responseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75161,8 +75161,8 @@ var require_test_transport = __commonJS({ stream.notifyMessage(msg); try { yield delay(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } } @@ -76225,8 +76225,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76322,8 +76322,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76354,18 +76354,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76633,8 +76633,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76835,22 +76835,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76905,15 +76905,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -76921,8 +76921,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76984,10 +76984,10 @@ var require_cache3 = __commonJS({ } core14.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else { @@ -77000,8 +77000,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77046,8 +77046,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core14.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core14.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError2(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core14.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77066,10 +77066,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77084,8 +77084,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -80413,8 +80413,8 @@ var require_util11 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } function maskSecretUrls(body) { @@ -80507,8 +80507,8 @@ var require_artifact_twirp_client2 = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -80539,18 +80539,18 @@ var require_artifact_twirp_client2 = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -80821,11 +80821,11 @@ var require_blob_upload = __commonJS({ blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options), chunkTimer((0, config_1.getUploadChunkTimeout)()) ]); - } catch (error4) { - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + } catch (error3) { + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } - throw error4; + throw error3; } finally { abortController.abort(); } @@ -81846,9 +81846,9 @@ var require_async = __commonJS({ invokeCallback(callback, err && (err instanceof Error || err.message) ? err : new Error(err)); }); } - function invokeCallback(callback, error4, value) { + function invokeCallback(callback, error3, value) { try { - callback(error4, value); + callback(error3, value); } catch (err) { setImmediate$1((e) => { throw e; @@ -83154,10 +83154,10 @@ var require_async = __commonJS({ function reflect(fn) { var _fn = wrapAsync(fn); return initialParams(function reflectOn(args, reflectCallback) { - args.push((error4, ...cbArgs) => { + args.push((error3, ...cbArgs) => { let retVal = {}; - if (error4) { - retVal.error = error4; + if (error3) { + retVal.error = error3; } if (cbArgs.length > 0) { var value = cbArgs; @@ -83313,13 +83313,13 @@ var require_async = __commonJS({ var timer; function timeoutCallback() { var name = asyncFn.name || "anonymous"; - var error4 = new Error('Callback function "' + name + '" timed out.'); - error4.code = "ETIMEDOUT"; + var error3 = new Error('Callback function "' + name + '" timed out.'); + error3.code = "ETIMEDOUT"; if (info7) { - error4.info = info7; + error3.info = info7; } timedOut = true; - callback(error4); + callback(error3); } args.push((...cbArgs) => { if (!timedOut) { @@ -83362,7 +83362,7 @@ var require_async = __commonJS({ return callback[PROMISE_SYMBOL]; } function tryEach(tasks, callback) { - var error4 = null; + var error3 = null; var result; return eachSeries$1(tasks, (task, taskCb) => { wrapAsync(task)((err, ...args) => { @@ -83372,10 +83372,10 @@ var require_async = __commonJS({ } else { result = args; } - error4 = err; + error3 = err; taskCb(err ? null : {}); }); - }, () => callback(error4, result)); + }, () => callback(error3, result)); } var tryEach$1 = awaitify(tryEach); function unmemoize(fn) { @@ -89855,19 +89855,19 @@ var require_from = __commonJS({ next(); } }; - readable._destroy = function(error4, cb) { + readable._destroy = function(error3, cb) { PromisePrototypeThen( - close(error4), - () => process2.nextTick(cb, error4), + close(error3), + () => process2.nextTick(cb, error3), // nextTick is here in case cb throws - (e) => process2.nextTick(cb, e || error4) + (e) => process2.nextTick(cb, e || error3) ); }; - async function close(error4) { - const hadError = error4 !== void 0 && error4 !== null; + async function close(error3) { + const hadError = error3 !== void 0 && error3 !== null; const hasThrow = typeof iterator.throw === "function"; if (hadError && hasThrow) { - const { value, done } = await iterator.throw(error4); + const { value, done } = await iterator.throw(error3); await value; if (done) { return; @@ -90075,12 +90075,12 @@ var require_readable3 = __commonJS({ this.destroy(err); }; Readable.prototype[SymbolAsyncDispose] = function() { - let error4; + let error3; if (!this.destroyed) { - error4 = this.readableEnded ? null : new AbortError(); - this.destroy(error4); + error3 = this.readableEnded ? null : new AbortError(); + this.destroy(error3); } - return new Promise2((resolve2, reject) => eos(this, (err) => err && err !== error4 ? reject(err) : resolve2(null))); + return new Promise2((resolve2, reject) => eos(this, (err) => err && err !== error3 ? reject(err) : resolve2(null))); }; Readable.prototype.push = function(chunk, encoding) { return readableAddChunk(this, chunk, encoding, false); @@ -90633,14 +90633,14 @@ var require_readable3 = __commonJS({ } } stream.on("readable", next); - let error4; + let error3; const cleanup = eos( stream, { writable: false }, (err) => { - error4 = err ? aggregateTwoErrors(error4, err) : null; + error3 = err ? aggregateTwoErrors(error3, err) : null; callback(); callback = nop; } @@ -90650,19 +90650,19 @@ var require_readable3 = __commonJS({ const chunk = stream.destroyed ? null : stream.read(); if (chunk !== null) { yield chunk; - } else if (error4) { - throw error4; - } else if (error4 === null) { + } else if (error3) { + throw error3; + } else if (error3 === null) { return; } else { await new Promise2(next); } } } catch (err) { - error4 = aggregateTwoErrors(error4, err); - throw error4; + error3 = aggregateTwoErrors(error3, err); + throw error3; } finally { - if ((error4 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error4 === void 0 || stream._readableState.autoDestroy)) { + if ((error3 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error3 === void 0 || stream._readableState.autoDestroy)) { destroyImpl.destroyer(stream, null); } else { stream.off("readable", next); @@ -92155,11 +92155,11 @@ var require_pipeline3 = __commonJS({ yield* Readable.prototype[SymbolAsyncIterator].call(val2); } async function pumpToNode(iterable, writable, finish, { end }) { - let error4; + let error3; let onresolve = null; const resume = (err) => { if (err) { - error4 = err; + error3 = err; } if (onresolve) { const callback = onresolve; @@ -92168,12 +92168,12 @@ var require_pipeline3 = __commonJS({ } }; const wait = () => new Promise2((resolve2, reject) => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { onresolve = () => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve2(); } @@ -92203,7 +92203,7 @@ var require_pipeline3 = __commonJS({ } finish(); } catch (err) { - finish(error4 !== err ? aggregateTwoErrors(error4, err) : err); + finish(error3 !== err ? aggregateTwoErrors(error3, err) : err); } finally { cleanup(); writable.off("drain", resume); @@ -92257,7 +92257,7 @@ var require_pipeline3 = __commonJS({ if (outerSignal) { disposable = addAbortListener(outerSignal, abort); } - let error4; + let error3; let value; const destroys = []; let finishCount = 0; @@ -92266,23 +92266,23 @@ var require_pipeline3 = __commonJS({ } function finishImpl(err, final) { var _disposable; - if (err && (!error4 || error4.code === "ERR_STREAM_PREMATURE_CLOSE")) { - error4 = err; + if (err && (!error3 || error3.code === "ERR_STREAM_PREMATURE_CLOSE")) { + error3 = err; } - if (!error4 && !final) { + if (!error3 && !final) { return; } while (destroys.length) { - destroys.shift()(error4); + destroys.shift()(error3); } ; (_disposable = disposable) === null || _disposable === void 0 ? void 0 : _disposable[SymbolDispose](); ac.abort(); if (final) { - if (!error4) { + if (!error3) { lastStreamCleanup.forEach((fn) => fn()); } - process2.nextTick(callback, error4, value); + process2.nextTick(callback, error3, value); } } let ret; @@ -102624,18 +102624,18 @@ var require_zip_archive_output_stream = __commonJS({ ZipArchiveOutputStream.prototype._smartStream = function(ae, callback) { var deflate = ae.getMethod() === constants.METHOD_DEFLATED; var process2 = deflate ? new DeflateCRC32Stream(this.options.zlib) : new CRC32Stream(); - var error4 = null; + var error3 = null; function handleStuff() { var digest = process2.digest().readUInt32BE(0); ae.setCrc(digest); ae.setSize(process2.size()); ae.setCompressedSize(process2.size(true)); this._afterAppend(ae); - callback(error4, ae); + callback(error3, ae); } process2.once("end", handleStuff.bind(this)); process2.once("error", function(err) { - error4 = err; + error3 = err; }); process2.pipe(this, { end: false }); return process2; @@ -104001,11 +104001,11 @@ var require_streamx = __commonJS({ } [asyncIterator]() { const stream = this; - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("readable", onreadable); this.on("close", onclose); @@ -104037,7 +104037,7 @@ var require_streamx = __commonJS({ } function ondata(data) { if (promiseReject === null) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else if (data === null && (stream._duplexState & READ_DONE) === 0) promiseReject(STREAM_DESTROYED); else promiseResolve({ value: data, done: data === null }); promiseReject = promiseResolve = null; @@ -104211,7 +104211,7 @@ var require_streamx = __commonJS({ if (all.length < 2) throw new Error("Pipeline requires at least 2 streams"); let src = all[0]; let dest = null; - let error4 = null; + let error3 = null; for (let i = 1; i < all.length; i++) { dest = all[i]; if (isStreamx(src)) { @@ -104226,14 +104226,14 @@ var require_streamx = __commonJS({ let fin = false; const autoDestroy = isStreamx(dest) || !!(dest._writableState && dest._writableState.autoDestroy); dest.on("error", (err) => { - if (error4 === null) error4 = err; + if (error3 === null) error3 = err; }); dest.on("finish", () => { fin = true; - if (!autoDestroy) done(error4); + if (!autoDestroy) done(error3); }); if (autoDestroy) { - dest.on("close", () => done(error4 || (fin ? null : PREMATURE_CLOSE))); + dest.on("close", () => done(error3 || (fin ? null : PREMATURE_CLOSE))); } } return dest; @@ -104246,8 +104246,8 @@ var require_streamx = __commonJS({ } } function onerror(err) { - if (!err || error4) return; - error4 = err; + if (!err || error3) return; + error3 = err; for (const s of all) { s.destroy(err); } @@ -104813,7 +104813,7 @@ var require_extract = __commonJS({ cb(null); } [Symbol.asyncIterator]() { - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; let entryStream = null; @@ -104821,7 +104821,7 @@ var require_extract = __commonJS({ const extract2 = this; this.on("entry", onentry); this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("close", onclose); return { @@ -104845,8 +104845,8 @@ var require_extract = __commonJS({ cb(err); } function onnext(resolve2, reject) { - if (error4) { - return reject(error4); + if (error3) { + return reject(error3); } if (entryStream) { resolve2({ value: entryStream, done: false }); @@ -104872,9 +104872,9 @@ var require_extract = __commonJS({ } } function onclose() { - consumeCallback(error4); + consumeCallback(error3); if (!promiseResolve) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else promiseResolve({ value: void 0, done: true }); promiseResolve = promiseReject = null; } @@ -105770,18 +105770,18 @@ var require_zip2 = __commonJS({ return zipUploadStream; }); } - var zipErrorCallback = (error4) => { + var zipErrorCallback = (error3) => { core14.error("An error has occurred while creating the zip file for upload"); - core14.info(error4); + core14.info(error3); throw new Error("An error has occurred during zip creation for the artifact"); }; - var zipWarningCallback = (error4) => { - if (error4.code === "ENOENT") { + var zipWarningCallback = (error3) => { + if (error3.code === "ENOENT") { core14.warning("ENOENT warning during artifact zip creation. No such file or directory"); - core14.info(error4); + core14.info(error3); } else { - core14.warning(`A non-blocking warning has occurred during artifact zip creation: ${error4.code}`); - core14.info(error4); + core14.warning(`A non-blocking warning has occurred during artifact zip creation: ${error3.code}`); + core14.info(error3); } }; var zipFinishCallback = () => { @@ -107588,8 +107588,8 @@ var require_parser_stream = __commonJS({ this.unzipStream.on("entry", function(entry) { self2.push(entry); }); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(ParserStream, Transform); @@ -107729,8 +107729,8 @@ var require_extract2 = __commonJS({ this.createdDirectories = {}; var self2 = this; this.unzipStream.on("entry", this._processEntry.bind(this)); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(Extract, Transform); @@ -107764,8 +107764,8 @@ var require_extract2 = __commonJS({ self2.unfinishedEntries--; self2._notifyAwaiter(); }); - pipedStream.on("error", function(error4) { - self2.emit("error", error4); + pipedStream.on("error", function(error3) { + self2.emit("error", error3); }); entry.pipe(pipedStream); }; @@ -107900,11 +107900,11 @@ var require_download_artifact = __commonJS({ try { yield promises_1.default.access(path2); return true; - } catch (error4) { - if (error4.code === "ENOENT") { + } catch (error3) { + if (error3.code === "ENOENT") { return false; } else { - throw error4; + throw error3; } } }); @@ -107915,9 +107915,9 @@ var require_download_artifact = __commonJS({ while (retryCount < 5) { try { return yield streamExtractExternal(url, directory); - } catch (error4) { + } catch (error3) { retryCount++; - core14.debug(`Failed to download artifact after ${retryCount} retries due to ${error4.message}. Retrying in 5 seconds...`); + core14.debug(`Failed to download artifact after ${retryCount} retries due to ${error3.message}. Retrying in 5 seconds...`); yield new Promise((resolve2) => setTimeout(resolve2, 5e3)); } } @@ -107946,10 +107946,10 @@ var require_download_artifact = __commonJS({ const extractStream = passThrough; extractStream.on("data", () => { timer.refresh(); - }).on("error", (error4) => { - core14.debug(`response.message: Artifact download failed: ${error4.message}`); + }).on("error", (error3) => { + core14.debug(`response.message: Artifact download failed: ${error3.message}`); clearTimeout(timer); - reject(error4); + reject(error3); }).pipe(unzip_stream_1.default.Extract({ path: directory })).on("close", () => { clearTimeout(timer); if (hashStream) { @@ -107958,8 +107958,8 @@ var require_download_artifact = __commonJS({ core14.info(`SHA256 digest of downloaded artifact is ${sha256Digest}`); } resolve2({ sha256Digest: `sha256:${sha256Digest}` }); - }).on("error", (error4) => { - reject(error4); + }).on("error", (error3) => { + reject(error3); }); }); }); @@ -107998,8 +107998,8 @@ var require_download_artifact = __commonJS({ core14.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -108041,8 +108041,8 @@ Are you trying to download from a different run? Try specifying a github-token w core14.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -108140,9 +108140,9 @@ var require_dist_node16 = __commonJS({ return request(options).then((response) => { octokit.log.info(`${requestOptions.method} ${path2} - ${response.status} in ${Date.now() - start}ms`); return response; - }).catch((error4) => { - octokit.log.info(`${requestOptions.method} ${path2} - ${error4.status} in ${Date.now() - start}ms`); - throw error4; + }).catch((error3) => { + octokit.log.info(`${requestOptions.method} ${path2} - ${error3.status} in ${Date.now() - start}ms`); + throw error3; }); }); } @@ -108160,22 +108160,22 @@ var require_dist_node17 = __commonJS({ return ex && typeof ex === "object" && "default" in ex ? ex["default"] : ex; } var Bottleneck = _interopDefault(require_light()); - async function errorRequest(octokit, state, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(octokit, state, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } async function wrapRequest(state, request, options) { const limiter = new Bottleneck(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -108197,12 +108197,12 @@ var require_dist_node17 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -108693,13 +108693,13 @@ var require_client2 = __commonJS({ throw new errors_1.GHESNotSupportedError(); } return (0, upload_artifact_1.uploadArtifact)(name, files, rootDirectory, options); - } catch (error4) { - (0, core_1.warning)(`Artifact upload failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Artifact upload failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions is operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -108714,13 +108714,13 @@ If the error persists, please check whether Actions is operating normally at [ht return (0, download_artifact_1.downloadArtifactPublic)(artifactId, repositoryOwner, repositoryName, token, downloadOptions); } return (0, download_artifact_1.downloadArtifactInternal)(artifactId, options); - } catch (error4) { - (0, core_1.warning)(`Download Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Download Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -108735,13 +108735,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, list_artifacts_1.listArtifactsPublic)(workflowRunId, repositoryOwner, repositoryName, token, options === null || options === void 0 ? void 0 : options.latest); } return (0, list_artifacts_1.listArtifactsInternal)(options === null || options === void 0 ? void 0 : options.latest); - } catch (error4) { - (0, core_1.warning)(`Listing Artifacts failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Listing Artifacts failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -108756,13 +108756,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, get_artifact_1.getArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, get_artifact_1.getArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Get Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Get Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -108777,13 +108777,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, delete_artifact_1.deleteArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, delete_artifact_1.deleteArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Delete Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Delete Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -109283,14 +109283,14 @@ var require_tmp = __commonJS({ options.template = _getRelativePathSync("template", options.template, tmpDir); return options; } - function _isEBADF(error4) { - return _isExpectedError(error4, -EBADF, "EBADF"); + function _isEBADF(error3) { + return _isExpectedError(error3, -EBADF, "EBADF"); } - function _isENOENT(error4) { - return _isExpectedError(error4, -ENOENT, "ENOENT"); + function _isENOENT(error3) { + return _isExpectedError(error3, -ENOENT, "ENOENT"); } - function _isExpectedError(error4, errno, code) { - return IS_WIN32 ? error4.code === code : error4.code === code && error4.errno === errno; + function _isExpectedError(error3, errno, code) { + return IS_WIN32 ? error3.code === code : error3.code === code && error3.errno === errno; } function setGracefulCleanup() { _gracefulCleanup = true; @@ -111002,9 +111002,9 @@ var require_upload_gzip = __commonJS({ const size = (yield stat(tempFilePath)).size; resolve2(size); })); - outputStream.on("error", (error4) => { - console.log(error4); - reject(error4); + outputStream.on("error", (error3) => { + console.log(error3); + reject(error3); }); }); }); @@ -111129,9 +111129,9 @@ var require_requestUtils2 = __commonJS({ } isRetryable = (0, utils_1.isRetryableStatusCode)(statusCode); errorMessage = `Artifact service responded with ${statusCode}`; - } catch (error4) { + } catch (error3) { isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { core14.info(`${name} - Error is not retryable`); @@ -111497,9 +111497,9 @@ var require_upload_http_client = __commonJS({ let response; try { response = yield uploadChunkRequest(); - } catch (error4) { + } catch (error3) { core14.info(`An error has been caught http-client index ${httpClientIndex}, retrying the upload`); - console.log(error4); + console.log(error3); if (incrementAndCheckRetryLimit()) { return false; } @@ -111688,8 +111688,8 @@ var require_download_http_client = __commonJS({ } this.statusReporter.incrementProcessedCount(); } - }))).catch((error4) => { - throw new Error(`Unable to download the artifact: ${error4}`); + }))).catch((error3) => { + throw new Error(`Unable to download the artifact: ${error3}`); }).finally(() => { this.statusReporter.stop(); this.downloadHttpManager.disposeAndReplaceAllClients(); @@ -111754,9 +111754,9 @@ var require_download_http_client = __commonJS({ let response; try { response = yield makeDownloadRequest(); - } catch (error4) { + } catch (error3) { core14.info("An error occurred while attempting to download a file"); - console.log(error4); + console.log(error3); yield backOff(); continue; } @@ -111770,7 +111770,7 @@ var require_download_http_client = __commonJS({ } else { forceRetry = true; } - } catch (error4) { + } catch (error3) { forceRetry = true; } } @@ -111796,31 +111796,31 @@ var require_download_http_client = __commonJS({ yield new Promise((resolve2, reject) => { if (isGzip) { const gunzip = zlib.createGunzip(); - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core14.info(`An error occurred while attempting to read the response stream`); gunzip.close(); destinationStream.close(); - reject(error4); - }).pipe(gunzip).on("error", (error4) => { + reject(error3); + }).pipe(gunzip).on("error", (error3) => { core14.info(`An error occurred while attempting to decompress the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve2(); - }).on("error", (error4) => { + }).on("error", (error3) => { core14.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } else { - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core14.info(`An error occurred while attempting to read the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve2(); - }).on("error", (error4) => { + }).on("error", (error3) => { core14.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } }); @@ -114827,7 +114827,7 @@ var require_debug2 = __commonJS({ if (!debug4) { try { debug4 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug4 !== "function") { debug4 = function() { @@ -114860,8 +114860,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -114935,9 +114935,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -115104,10 +115104,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -115306,12 +115306,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -116461,14 +116461,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs2.lstat(itemPath, { bigint: true }) : await fs2.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs2.lstat(itemPath, { bigint: true }) : await fs2.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs2.readdir(itemPath) : await fs2.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs2.readdir(itemPath) : await fs2.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -116479,13 +116479,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -119170,8 +119170,8 @@ var ConfigurationError = class extends Error { super(message); } }; -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } // src/actions-util.ts @@ -119787,9 +119787,9 @@ async function runWrapper() { } ); } - } catch (error4) { + } catch (error3) { logger.warning( - `start-proxy post-action step failed: ${getErrorMessage(error4)}` + `start-proxy post-action step failed: ${getErrorMessage(error3)}` ); } } diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index cf0fbd92d..3b6d4deca 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug5("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve2(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug5; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning6(message, properties = {}) { + exports2.error = error3; + function warning7(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning6; + exports2.warning = warning7; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -24795,10 +24795,10 @@ var require_util8 = __commonJS({ rval = api.setItem(id, obj); } if (typeof rval !== "undefined" && rval.rval !== true) { - var error4 = new Error(rval.error.message); - error4.id = rval.error.id; - error4.name = rval.error.name; - throw error4; + var error3 = new Error(rval.error.message); + error3.id = rval.error.id; + error3.name = rval.error.name; + throw error3; } }; var _getStorageObject = function(api, id) { @@ -24809,10 +24809,10 @@ var require_util8 = __commonJS({ if (api.init) { if (rval.rval === null) { if (rval.error) { - var error4 = new Error(rval.error.message); - error4.id = rval.error.id; - error4.name = rval.error.name; - throw error4; + var error3 = new Error(rval.error.message); + error3.id = rval.error.id; + error3.name = rval.error.name; + throw error3; } rval = null; } else { @@ -26476,11 +26476,11 @@ var require_asn1 = __commonJS({ }; function _checkBufferLength(bytes, remaining, n) { if (n > remaining) { - var error4 = new Error("Too few bytes to parse DER."); - error4.available = bytes.length(); - error4.remaining = remaining; - error4.requested = n; - throw error4; + var error3 = new Error("Too few bytes to parse DER."); + error3.available = bytes.length(); + error3.remaining = remaining; + error3.requested = n; + throw error3; } } var _getValueLength = function(bytes, remaining) { @@ -26533,10 +26533,10 @@ var require_asn1 = __commonJS({ var byteCount = bytes.length(); var value = _fromDer(bytes, bytes.length(), 0, options); if (options.parseAllBytes && bytes.length() !== 0) { - var error4 = new Error("Unparsed DER bytes remain after ASN.1 parsing."); - error4.byteCount = byteCount; - error4.remaining = bytes.length(); - throw error4; + var error3 = new Error("Unparsed DER bytes remain after ASN.1 parsing."); + error3.byteCount = byteCount; + error3.remaining = bytes.length(); + throw error3; } return value; }; @@ -26552,11 +26552,11 @@ var require_asn1 = __commonJS({ remaining -= start - bytes.length(); if (length !== void 0 && length > remaining) { if (options.strict) { - var error4 = new Error("Too few bytes to read ASN.1 value."); - error4.available = bytes.length(); - error4.remaining = remaining; - error4.requested = length; - throw error4; + var error3 = new Error("Too few bytes to read ASN.1 value."); + error3.available = bytes.length(); + error3.remaining = remaining; + error3.requested = length; + throw error3; } length = remaining; } @@ -26880,9 +26880,9 @@ var require_asn1 = __commonJS({ if (x >= -2147483648 && x < 2147483648) { return rval.putSignedInt(x, 32); } - var error4 = new Error("Integer too large; max is 32-bits."); - error4.integer = x; - throw error4; + var error3 = new Error("Integer too large; max is 32-bits."); + error3.integer = x; + throw error3; }; asn1.derToInteger = function(bytes) { if (typeof bytes === "string") { @@ -30475,10 +30475,10 @@ var require_pkcs1 = __commonJS({ var keyLength = Math.ceil(key.n.bitLength() / 8); var maxLength = keyLength - 2 * md.digestLength - 2; if (message.length > maxLength) { - var error4 = new Error("RSAES-OAEP input message length is too long."); - error4.length = message.length; - error4.maxLength = maxLength; - throw error4; + var error3 = new Error("RSAES-OAEP input message length is too long."); + error3.length = message.length; + error3.maxLength = maxLength; + throw error3; } if (!label) { label = ""; @@ -30494,10 +30494,10 @@ var require_pkcs1 = __commonJS({ if (!seed) { seed = forge.random.getBytes(md.digestLength); } else if (seed.length !== md.digestLength) { - var error4 = new Error("Invalid RSAES-OAEP seed. The seed length must match the digest length."); - error4.seedLength = seed.length; - error4.digestLength = md.digestLength; - throw error4; + var error3 = new Error("Invalid RSAES-OAEP seed. The seed length must match the digest length."); + error3.seedLength = seed.length; + error3.digestLength = md.digestLength; + throw error3; } var dbMask = rsa_mgf1(seed, keyLength - md.digestLength - 1, mgf1Md); var maskedDB = forge.util.xorBytes(DB, dbMask, DB.length); @@ -30521,10 +30521,10 @@ var require_pkcs1 = __commonJS({ } var keyLength = Math.ceil(key.n.bitLength() / 8); if (em.length !== keyLength) { - var error4 = new Error("RSAES-OAEP encoded message length is invalid."); - error4.length = em.length; - error4.expectedLength = keyLength; - throw error4; + var error3 = new Error("RSAES-OAEP encoded message length is invalid."); + error3.length = em.length; + error3.expectedLength = keyLength; + throw error3; } if (md === void 0) { md = forge.md.sha1.create(); @@ -30550,9 +30550,9 @@ var require_pkcs1 = __commonJS({ var dbMask = rsa_mgf1(seed, keyLength - md.digestLength - 1, mgf1Md); var db = forge.util.xorBytes(maskedDB, dbMask, maskedDB.length); var lHashPrime = db.substring(0, md.digestLength); - var error4 = y !== "\0"; + var error3 = y !== "\0"; for (var i = 0; i < md.digestLength; ++i) { - error4 |= lHash.charAt(i) !== lHashPrime.charAt(i); + error3 |= lHash.charAt(i) !== lHashPrime.charAt(i); } var in_ps = 1; var index = md.digestLength; @@ -30560,11 +30560,11 @@ var require_pkcs1 = __commonJS({ var code = db.charCodeAt(j); var is_0 = code & 1 ^ 1; var error_mask = in_ps ? 65534 : 0; - error4 |= code & error_mask; + error3 |= code & error_mask; in_ps = in_ps & is_0; index += in_ps; } - if (error4 || db.charCodeAt(index) !== 1) { + if (error3 || db.charCodeAt(index) !== 1) { throw new Error("Invalid RSAES-OAEP padding."); } return db.substring(index + 1); @@ -30978,9 +30978,9 @@ var require_rsa = __commonJS({ if (md.algorithm in pki2.oids) { oid = pki2.oids[md.algorithm]; } else { - var error4 = new Error("Unknown message digest algorithm."); - error4.algorithm = md.algorithm; - throw error4; + var error3 = new Error("Unknown message digest algorithm."); + error3.algorithm = md.algorithm; + throw error3; } var oidBytes = asn1.oidToDer(oid).getBytes(); var digestInfo = asn1.create( @@ -31076,10 +31076,10 @@ var require_rsa = __commonJS({ pki2.rsa.decrypt = function(ed, key, pub, ml) { var k = Math.ceil(key.n.bitLength() / 8); if (ed.length !== k) { - var error4 = new Error("Encrypted message length is invalid."); - error4.length = ed.length; - error4.expected = k; - throw error4; + var error3 = new Error("Encrypted message length is invalid."); + error3.length = ed.length; + error3.expected = k; + throw error3; } var y = new BigInteger(forge.util.createBuffer(ed).toHex(), 16); if (y.compareTo(key.n) >= 0) { @@ -31451,19 +31451,19 @@ var require_rsa = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, digestInfoValidator, capture, errors)) { - var error4 = new Error( + var error3 = new Error( "ASN.1 object does not contain a valid RSASSA-PKCS1-v1_5 DigestInfo value." ); - error4.errors = errors; - throw error4; + error3.errors = errors; + throw error3; } var oid = asn1.derToOid(capture.algorithmIdentifier); if (!(oid === forge.oids.md2 || oid === forge.oids.md5 || oid === forge.oids.sha1 || oid === forge.oids.sha224 || oid === forge.oids.sha256 || oid === forge.oids.sha384 || oid === forge.oids.sha512 || oid === forge.oids["sha512-224"] || oid === forge.oids["sha512-256"])) { - var error4 = new Error( + var error3 = new Error( "Unknown RSASSA-PKCS1-v1_5 DigestAlgorithm identifier." ); - error4.oid = oid; - throw error4; + error3.oid = oid; + throw error3; } if (oid === forge.oids.md2 || oid === forge.oids.md5) { if (!("parameters" in capture)) { @@ -31579,9 +31579,9 @@ var require_rsa = __commonJS({ capture = {}; errors = []; if (!asn1.validate(obj, rsaPrivateKeyValidator, capture, errors)) { - var error4 = new Error("Cannot read private key. ASN.1 object does not contain an RSAPrivateKey."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read private key. ASN.1 object does not contain an RSAPrivateKey."); + error3.errors = errors; + throw error3; } var n, e, d, p, q, dP, dQ, qInv; n = forge.util.createBuffer(capture.privateKeyModulus).toHex(); @@ -31676,17 +31676,17 @@ var require_rsa = __commonJS({ if (asn1.validate(obj, publicKeyValidator, capture, errors)) { var oid = asn1.derToOid(capture.publicKeyOid); if (oid !== pki2.oids.rsaEncryption) { - var error4 = new Error("Cannot read public key. Unknown OID."); - error4.oid = oid; - throw error4; + var error3 = new Error("Cannot read public key. Unknown OID."); + error3.oid = oid; + throw error3; } obj = capture.rsaPublicKey; } errors = []; if (!asn1.validate(obj, rsaPublicKeyValidator, capture, errors)) { - var error4 = new Error("Cannot read public key. ASN.1 object does not contain an RSAPublicKey."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read public key. ASN.1 object does not contain an RSAPublicKey."); + error3.errors = errors; + throw error3; } var n = forge.util.createBuffer(capture.publicKeyModulus).toHex(); var e = forge.util.createBuffer(capture.publicKeyExponent).toHex(); @@ -31737,10 +31737,10 @@ var require_rsa = __commonJS({ var eb = forge.util.createBuffer(); var k = Math.ceil(key.n.bitLength() / 8); if (m.length > k - 11) { - var error4 = new Error("Message is too long for PKCS#1 v1.5 padding."); - error4.length = m.length; - error4.max = k - 11; - throw error4; + var error3 = new Error("Message is too long for PKCS#1 v1.5 padding."); + error3.length = m.length; + error3.max = k - 11; + throw error3; } eb.putByte(0); eb.putByte(bt); @@ -32134,9 +32134,9 @@ var require_pbe = __commonJS({ cipherFn = forge.des.createEncryptionCipher; break; default: - var error4 = new Error("Cannot encrypt private key. Unknown encryption algorithm."); - error4.algorithm = options.algorithm; - throw error4; + var error3 = new Error("Cannot encrypt private key. Unknown encryption algorithm."); + error3.algorithm = options.algorithm; + throw error3; } var prfAlgorithm = "hmacWith" + options.prfAlgorithm.toUpperCase(); var md = prfAlgorithmToMessageDigest(prfAlgorithm); @@ -32226,9 +32226,9 @@ var require_pbe = __commonJS({ ] ); } else { - var error4 = new Error("Cannot encrypt private key. Unknown encryption algorithm."); - error4.algorithm = options.algorithm; - throw error4; + var error3 = new Error("Cannot encrypt private key. Unknown encryption algorithm."); + error3.algorithm = options.algorithm; + throw error3; } var rval = asn1.create(asn1.Class.UNIVERSAL, asn1.Type.SEQUENCE, true, [ // encryptionAlgorithm @@ -32248,9 +32248,9 @@ var require_pbe = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, encryptedPrivateKeyValidator, capture, errors)) { - var error4 = new Error("Cannot read encrypted private key. ASN.1 object is not a supported EncryptedPrivateKeyInfo."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read encrypted private key. ASN.1 object is not a supported EncryptedPrivateKeyInfo."); + error3.errors = errors; + throw error3; } var oid = asn1.derToOid(capture.encryptionOid); var cipher = pki2.pbe.getCipher(oid, capture.encryptionParams, password); @@ -32271,9 +32271,9 @@ var require_pbe = __commonJS({ pki2.encryptedPrivateKeyFromPem = function(pem) { var msg = forge.pem.decode(pem)[0]; if (msg.type !== "ENCRYPTED PRIVATE KEY") { - var error4 = new Error('Could not convert encrypted private key from PEM; PEM header type is "ENCRYPTED PRIVATE KEY".'); - error4.headerType = msg.type; - throw error4; + var error3 = new Error('Could not convert encrypted private key from PEM; PEM header type is "ENCRYPTED PRIVATE KEY".'); + error3.headerType = msg.type; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { throw new Error("Could not convert encrypted private key from PEM; PEM is encrypted."); @@ -32323,9 +32323,9 @@ var require_pbe = __commonJS({ cipherFn = forge.des.createEncryptionCipher; break; default: - var error4 = new Error('Could not encrypt RSA private key; unsupported encryption algorithm "' + options.algorithm + '".'); - error4.algorithm = options.algorithm; - throw error4; + var error3 = new Error('Could not encrypt RSA private key; unsupported encryption algorithm "' + options.algorithm + '".'); + error3.algorithm = options.algorithm; + throw error3; } var dk = forge.pbe.opensslDeriveBytes(password, iv.substr(0, 8), dkLen); var cipher = cipherFn(dk); @@ -32350,9 +32350,9 @@ var require_pbe = __commonJS({ var rval = null; var msg = forge.pem.decode(pem)[0]; if (msg.type !== "ENCRYPTED PRIVATE KEY" && msg.type !== "PRIVATE KEY" && msg.type !== "RSA PRIVATE KEY") { - var error4 = new Error('Could not convert private key from PEM; PEM header type is not "ENCRYPTED PRIVATE KEY", "PRIVATE KEY", or "RSA PRIVATE KEY".'); - error4.headerType = error4; - throw error4; + var error3 = new Error('Could not convert private key from PEM; PEM header type is not "ENCRYPTED PRIVATE KEY", "PRIVATE KEY", or "RSA PRIVATE KEY".'); + error3.headerType = error3; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { var dkLen; @@ -32397,9 +32397,9 @@ var require_pbe = __commonJS({ }; break; default: - var error4 = new Error('Could not decrypt private key; unsupported encryption algorithm "' + msg.dekInfo.algorithm + '".'); - error4.algorithm = msg.dekInfo.algorithm; - throw error4; + var error3 = new Error('Could not decrypt private key; unsupported encryption algorithm "' + msg.dekInfo.algorithm + '".'); + error3.algorithm = msg.dekInfo.algorithm; + throw error3; } var iv = forge.util.hexToBytes(msg.dekInfo.parameters); var dk = forge.pbe.opensslDeriveBytes(password, iv.substr(0, 8), dkLen); @@ -32498,43 +32498,43 @@ var require_pbe = __commonJS({ case pki2.oids["pbewithSHAAnd40BitRC2-CBC"]: return pki2.pbe.getCipherForPKCS12PBE(oid, params, password); default: - var error4 = new Error("Cannot read encrypted PBE data block. Unsupported OID."); - error4.oid = oid; - error4.supportedOids = [ + var error3 = new Error("Cannot read encrypted PBE data block. Unsupported OID."); + error3.oid = oid; + error3.supportedOids = [ "pkcs5PBES2", "pbeWithSHAAnd3-KeyTripleDES-CBC", "pbewithSHAAnd40BitRC2-CBC" ]; - throw error4; + throw error3; } }; pki2.pbe.getCipherForPBES2 = function(oid, params, password) { var capture = {}; var errors = []; if (!asn1.validate(params, PBES2AlgorithmsValidator, capture, errors)) { - var error4 = new Error("Cannot read password-based-encryption algorithm parameters. ASN.1 object is not a supported EncryptedPrivateKeyInfo."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read password-based-encryption algorithm parameters. ASN.1 object is not a supported EncryptedPrivateKeyInfo."); + error3.errors = errors; + throw error3; } oid = asn1.derToOid(capture.kdfOid); if (oid !== pki2.oids["pkcs5PBKDF2"]) { - var error4 = new Error("Cannot read encrypted private key. Unsupported key derivation function OID."); - error4.oid = oid; - error4.supportedOids = ["pkcs5PBKDF2"]; - throw error4; + var error3 = new Error("Cannot read encrypted private key. Unsupported key derivation function OID."); + error3.oid = oid; + error3.supportedOids = ["pkcs5PBKDF2"]; + throw error3; } oid = asn1.derToOid(capture.encOid); if (oid !== pki2.oids["aes128-CBC"] && oid !== pki2.oids["aes192-CBC"] && oid !== pki2.oids["aes256-CBC"] && oid !== pki2.oids["des-EDE3-CBC"] && oid !== pki2.oids["desCBC"]) { - var error4 = new Error("Cannot read encrypted private key. Unsupported encryption scheme OID."); - error4.oid = oid; - error4.supportedOids = [ + var error3 = new Error("Cannot read encrypted private key. Unsupported encryption scheme OID."); + error3.oid = oid; + error3.supportedOids = [ "aes128-CBC", "aes192-CBC", "aes256-CBC", "des-EDE3-CBC", "desCBC" ]; - throw error4; + throw error3; } var salt = capture.kdfSalt; var count = forge.util.createBuffer(capture.kdfIterationCount); @@ -32574,9 +32574,9 @@ var require_pbe = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(params, pkcs12PbeParamsValidator, capture, errors)) { - var error4 = new Error("Cannot read password-based-encryption algorithm parameters. ASN.1 object is not a supported EncryptedPrivateKeyInfo."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read password-based-encryption algorithm parameters. ASN.1 object is not a supported EncryptedPrivateKeyInfo."); + error3.errors = errors; + throw error3; } var salt = forge.util.createBuffer(capture.salt); var count = forge.util.createBuffer(capture.iterations); @@ -32598,9 +32598,9 @@ var require_pbe = __commonJS({ }; break; default: - var error4 = new Error("Cannot read PKCS #12 PBE data block. Unsupported OID."); - error4.oid = oid; - throw error4; + var error3 = new Error("Cannot read PKCS #12 PBE data block. Unsupported OID."); + error3.oid = oid; + throw error3; } var md = prfOidToMessageDigest(capture.prfOid); var key = pki2.pbe.generatePkcs12Key(password, salt, 1, count, dkLen, md); @@ -32634,16 +32634,16 @@ var require_pbe = __commonJS({ } else { prfAlgorithm = pki2.oids[asn1.derToOid(prfOid)]; if (!prfAlgorithm) { - var error4 = new Error("Unsupported PRF OID."); - error4.oid = prfOid; - error4.supported = [ + var error3 = new Error("Unsupported PRF OID."); + error3.oid = prfOid; + error3.supported = [ "hmacWithSHA1", "hmacWithSHA224", "hmacWithSHA256", "hmacWithSHA384", "hmacWithSHA512" ]; - throw error4; + throw error3; } } return prfAlgorithmToMessageDigest(prfAlgorithm); @@ -32660,16 +32660,16 @@ var require_pbe = __commonJS({ prfAlgorithm = prfAlgorithm.substr(8).toLowerCase(); break; default: - var error4 = new Error("Unsupported PRF algorithm."); - error4.algorithm = prfAlgorithm; - error4.supported = [ + var error3 = new Error("Unsupported PRF algorithm."); + error3.algorithm = prfAlgorithm; + error3.supported = [ "hmacWithSHA1", "hmacWithSHA224", "hmacWithSHA256", "hmacWithSHA384", "hmacWithSHA512" ]; - throw error4; + throw error3; } if (!factory || !(prfAlgorithm in factory)) { throw new Error("Unknown hash algorithm: " + prfAlgorithm); @@ -33666,9 +33666,9 @@ var require_x509 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, rsassaPssParameterValidator, capture, errors)) { - var error4 = new Error("Cannot read RSASSA-PSS parameter block."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read RSASSA-PSS parameter block."); + error3.errors = errors; + throw error3; } if (capture.hashOid !== void 0) { params.hash = params.hash || {}; @@ -33702,11 +33702,11 @@ var require_x509 = __commonJS({ case "RSASSA-PSS": return forge.md.sha256.create(); default: - var error4 = new Error( + var error3 = new Error( "Could not compute " + options.type + " digest. Unknown signature OID." ); - error4.signatureOid = options.signatureOid; - throw error4; + error3.signatureOid = options.signatureOid; + throw error3; } }; var _verifySignature = function(options) { @@ -33721,25 +33721,25 @@ var require_x509 = __commonJS({ var hash, mgf; hash = oids[cert.signatureParameters.mgf.hash.algorithmOid]; if (hash === void 0 || forge.md[hash] === void 0) { - var error4 = new Error("Unsupported MGF hash function."); - error4.oid = cert.signatureParameters.mgf.hash.algorithmOid; - error4.name = hash; - throw error4; + var error3 = new Error("Unsupported MGF hash function."); + error3.oid = cert.signatureParameters.mgf.hash.algorithmOid; + error3.name = hash; + throw error3; } mgf = oids[cert.signatureParameters.mgf.algorithmOid]; if (mgf === void 0 || forge.mgf[mgf] === void 0) { - var error4 = new Error("Unsupported MGF function."); - error4.oid = cert.signatureParameters.mgf.algorithmOid; - error4.name = mgf; - throw error4; + var error3 = new Error("Unsupported MGF function."); + error3.oid = cert.signatureParameters.mgf.algorithmOid; + error3.name = mgf; + throw error3; } mgf = forge.mgf[mgf].create(forge.md[hash].create()); hash = oids[cert.signatureParameters.hash.algorithmOid]; if (hash === void 0 || forge.md[hash] === void 0) { - var error4 = new Error("Unsupported RSASSA-PSS hash function."); - error4.oid = cert.signatureParameters.hash.algorithmOid; - error4.name = hash; - throw error4; + var error3 = new Error("Unsupported RSASSA-PSS hash function."); + error3.oid = cert.signatureParameters.hash.algorithmOid; + error3.name = hash; + throw error3; } scheme = forge.pss.create( forge.md[hash].create(), @@ -33757,11 +33757,11 @@ var require_x509 = __commonJS({ pki2.certificateFromPem = function(pem, computeHash, strict) { var msg = forge.pem.decode(pem)[0]; if (msg.type !== "CERTIFICATE" && msg.type !== "X509 CERTIFICATE" && msg.type !== "TRUSTED CERTIFICATE") { - var error4 = new Error( + var error3 = new Error( 'Could not convert certificate from PEM; PEM header type is not "CERTIFICATE", "X509 CERTIFICATE", or "TRUSTED CERTIFICATE".' ); - error4.headerType = msg.type; - throw error4; + error3.headerType = msg.type; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { throw new Error( @@ -33781,9 +33781,9 @@ var require_x509 = __commonJS({ pki2.publicKeyFromPem = function(pem) { var msg = forge.pem.decode(pem)[0]; if (msg.type !== "PUBLIC KEY" && msg.type !== "RSA PUBLIC KEY") { - var error4 = new Error('Could not convert public key from PEM; PEM header type is not "PUBLIC KEY" or "RSA PUBLIC KEY".'); - error4.headerType = msg.type; - throw error4; + var error3 = new Error('Could not convert public key from PEM; PEM header type is not "PUBLIC KEY" or "RSA PUBLIC KEY".'); + error3.headerType = msg.type; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { throw new Error("Could not convert public key from PEM; PEM is encrypted."); @@ -33839,9 +33839,9 @@ var require_x509 = __commonJS({ pki2.certificationRequestFromPem = function(pem, computeHash, strict) { var msg = forge.pem.decode(pem)[0]; if (msg.type !== "CERTIFICATE REQUEST") { - var error4 = new Error('Could not convert certification request from PEM; PEM header type is not "CERTIFICATE REQUEST".'); - error4.headerType = msg.type; - throw error4; + var error3 = new Error('Could not convert certification request from PEM; PEM header type is not "CERTIFICATE REQUEST".'); + error3.headerType = msg.type; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { throw new Error("Could not convert certification request from PEM; PEM is encrypted."); @@ -33934,9 +33934,9 @@ var require_x509 = __commonJS({ cert.md = md || forge.md.sha1.create(); var algorithmOid = oids[cert.md.algorithm + "WithRSAEncryption"]; if (!algorithmOid) { - var error4 = new Error("Could not compute certificate digest. Unknown message digest algorithm OID."); - error4.algorithm = cert.md.algorithm; - throw error4; + var error3 = new Error("Could not compute certificate digest. Unknown message digest algorithm OID."); + error3.algorithm = cert.md.algorithm; + throw error3; } cert.signatureOid = cert.siginfo.algorithmOid = algorithmOid; cert.tbsCertificate = pki2.getTBSCertificate(cert); @@ -33949,12 +33949,12 @@ var require_x509 = __commonJS({ if (!cert.issued(child)) { var issuer = child.issuer; var subject = cert.subject; - var error4 = new Error( + var error3 = new Error( "The parent certificate did not issue the given child certificate; the child certificate's issuer does not match the parent's subject." ); - error4.expectedIssuer = subject.attributes; - error4.actualIssuer = issuer.attributes; - throw error4; + error3.expectedIssuer = subject.attributes; + error3.actualIssuer = issuer.attributes; + throw error3; } var md = child.md; if (md === null) { @@ -34017,9 +34017,9 @@ var require_x509 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, x509CertificateValidator, capture, errors)) { - var error4 = new Error("Cannot read X.509 certificate. ASN.1 object is not an X509v3 Certificate."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read X.509 certificate. ASN.1 object is not an X509v3 Certificate."); + error3.errors = errors; + throw error3; } var oid = asn1.derToOid(capture.publicKeyOid); if (oid !== pki2.oids.rsaEncryption) { @@ -34234,9 +34234,9 @@ var require_x509 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, certificationRequestValidator, capture, errors)) { - var error4 = new Error("Cannot read PKCS#10 certificate request. ASN.1 object is not a PKCS#10 CertificationRequest."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read PKCS#10 certificate request. ASN.1 object is not a PKCS#10 CertificationRequest."); + error3.errors = errors; + throw error3; } var oid = asn1.derToOid(capture.publicKeyOid); if (oid !== pki2.oids.rsaEncryption) { @@ -34332,9 +34332,9 @@ var require_x509 = __commonJS({ csr.md = md || forge.md.sha1.create(); var algorithmOid = oids[csr.md.algorithm + "WithRSAEncryption"]; if (!algorithmOid) { - var error4 = new Error("Could not compute certification request digest. Unknown message digest algorithm OID."); - error4.algorithm = csr.md.algorithm; - throw error4; + var error3 = new Error("Could not compute certification request digest. Unknown message digest algorithm OID."); + error3.algorithm = csr.md.algorithm; + throw error3; } csr.signatureOid = csr.siginfo.algorithmOid = algorithmOid; csr.certificationRequestInfo = pki2.getCertificationRequestInfo(csr); @@ -34416,9 +34416,9 @@ var require_x509 = __commonJS({ if (attr.name && attr.name in pki2.oids) { attr.type = pki2.oids[attr.name]; } else { - var error4 = new Error("Attribute type not specified."); - error4.attribute = attr; - throw error4; + var error3 = new Error("Attribute type not specified."); + error3.attribute = attr; + throw error3; } } if (typeof attr.shortName === "undefined") { @@ -34439,9 +34439,9 @@ var require_x509 = __commonJS({ } } if (typeof attr.value === "undefined") { - var error4 = new Error("Attribute value not specified."); - error4.attribute = attr; - throw error4; + var error3 = new Error("Attribute value not specified."); + error3.attribute = attr; + throw error3; } } } @@ -34456,9 +34456,9 @@ var require_x509 = __commonJS({ if (e.name && e.name in pki2.oids) { e.id = pki2.oids[e.name]; } else { - var error4 = new Error("Extension ID not specified."); - error4.extension = e; - throw error4; + var error3 = new Error("Extension ID not specified."); + error3.extension = e; + throw error3; } } if (typeof e.value !== "undefined") { @@ -34621,11 +34621,11 @@ var require_x509 = __commonJS({ if (altName.type === 7 && altName.ip) { value = forge.util.bytesFromIP(altName.ip); if (value === null) { - var error4 = new Error( + var error3 = new Error( 'Extension "ip" value is not a valid IPv4 or IPv6 address.' ); - error4.extension = e; - throw error4; + error3.extension = e; + throw error3; } } else if (altName.type === 8) { if (altName.oid) { @@ -34707,11 +34707,11 @@ var require_x509 = __commonJS({ if (altName.type === 7 && altName.ip) { value = forge.util.bytesFromIP(altName.ip); if (value === null) { - var error4 = new Error( + var error3 = new Error( 'Extension "ip" value is not a valid IPv4 or IPv6 address.' ); - error4.extension = e; - throw error4; + error3.extension = e; + throw error3; } } else if (altName.type === 8) { if (altName.oid) { @@ -34736,9 +34736,9 @@ var require_x509 = __commonJS({ seq2.push(subSeq); } if (typeof e.value === "undefined") { - var error4 = new Error("Extension value not specified."); - error4.extension = e; - throw error4; + var error3 = new Error("Extension value not specified."); + error3.extension = e; + throw error3; } return e; } @@ -35175,7 +35175,7 @@ var require_x509 = __commonJS({ validityCheckDate = /* @__PURE__ */ new Date(); } var first = true; - var error4 = null; + var error3 = null; var depth = 0; do { var cert = chain.shift(); @@ -35183,7 +35183,7 @@ var require_x509 = __commonJS({ var selfSigned = false; if (validityCheckDate) { if (validityCheckDate < cert.validity.notBefore || validityCheckDate > cert.validity.notAfter) { - error4 = { + error3 = { message: "Certificate is not valid yet or has expired.", error: pki2.certificateError.certificate_expired, notBefore: cert.validity.notBefore, @@ -35194,7 +35194,7 @@ var require_x509 = __commonJS({ }; } } - if (error4 === null) { + if (error3 === null) { parent = chain[0] || caStore.getIssuer(cert); if (parent === null) { if (cert.isIssuer(cert)) { @@ -35216,74 +35216,74 @@ var require_x509 = __commonJS({ } } if (!verified) { - error4 = { + error3 = { message: "Certificate signature is invalid.", error: pki2.certificateError.bad_certificate }; } } - if (error4 === null && (!parent || selfSigned) && !caStore.hasCertificate(cert)) { - error4 = { + if (error3 === null && (!parent || selfSigned) && !caStore.hasCertificate(cert)) { + error3 = { message: "Certificate is not trusted.", error: pki2.certificateError.unknown_ca }; } } - if (error4 === null && parent && !cert.isIssuer(parent)) { - error4 = { + if (error3 === null && parent && !cert.isIssuer(parent)) { + error3 = { message: "Certificate issuer is invalid.", error: pki2.certificateError.bad_certificate }; } - if (error4 === null) { + if (error3 === null) { var se = { keyUsage: true, basicConstraints: true }; - for (var i = 0; error4 === null && i < cert.extensions.length; ++i) { + for (var i = 0; error3 === null && i < cert.extensions.length; ++i) { var ext = cert.extensions[i]; if (ext.critical && !(ext.name in se)) { - error4 = { + error3 = { message: "Certificate has an unsupported critical extension.", error: pki2.certificateError.unsupported_certificate }; } } } - if (error4 === null && (!first || chain.length === 0 && (!parent || selfSigned))) { + if (error3 === null && (!first || chain.length === 0 && (!parent || selfSigned))) { var bcExt = cert.getExtension("basicConstraints"); var keyUsageExt = cert.getExtension("keyUsage"); if (keyUsageExt !== null) { if (!keyUsageExt.keyCertSign || bcExt === null) { - error4 = { + error3 = { message: "Certificate keyUsage or basicConstraints conflict or indicate that the certificate is not a CA. If the certificate is the only one in the chain or isn't the first then the certificate must be a valid CA.", error: pki2.certificateError.bad_certificate }; } } - if (error4 === null && bcExt !== null && !bcExt.cA) { - error4 = { + if (error3 === null && bcExt !== null && !bcExt.cA) { + error3 = { message: "Certificate basicConstraints indicates the certificate is not a CA.", error: pki2.certificateError.bad_certificate }; } - if (error4 === null && keyUsageExt !== null && "pathLenConstraint" in bcExt) { + if (error3 === null && keyUsageExt !== null && "pathLenConstraint" in bcExt) { var pathLen = depth - 1; if (pathLen > bcExt.pathLenConstraint) { - error4 = { + error3 = { message: "Certificate basicConstraints pathLenConstraint violated.", error: pki2.certificateError.bad_certificate }; } } } - var vfd = error4 === null ? true : error4.error; + var vfd = error3 === null ? true : error3.error; var ret = options.verify ? options.verify(vfd, depth, certs) : vfd; if (ret === true) { - error4 = null; + error3 = null; } else { if (vfd === true) { - error4 = { + error3 = { message: "The application rejected the certificate.", error: pki2.certificateError.bad_certificate }; @@ -35291,16 +35291,16 @@ var require_x509 = __commonJS({ if (ret || ret === 0) { if (typeof ret === "object" && !forge.util.isArray(ret)) { if (ret.message) { - error4.message = ret.message; + error3.message = ret.message; } if (ret.error) { - error4.error = ret.error; + error3.error = ret.error; } } else if (typeof ret === "string") { - error4.error = ret; + error3.error = ret; } } - throw error4; + throw error3; } first = false; ++depth; @@ -35513,9 +35513,9 @@ var require_pkcs12 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, pfxValidator, capture, errors)) { - var error4 = new Error("Cannot read PKCS#12 PFX. ASN.1 object is not an PKCS#12 PFX."); - error4.errors = error4; - throw error4; + var error3 = new Error("Cannot read PKCS#12 PFX. ASN.1 object is not an PKCS#12 PFX."); + error3.errors = error3; + throw error3; } var pfx = { version: capture.version.charCodeAt(0), @@ -35605,14 +35605,14 @@ var require_pkcs12 = __commonJS({ } }; if (capture.version.charCodeAt(0) !== 3) { - var error4 = new Error("PKCS#12 PFX of version other than 3 not supported."); - error4.version = capture.version.charCodeAt(0); - throw error4; + var error3 = new Error("PKCS#12 PFX of version other than 3 not supported."); + error3.version = capture.version.charCodeAt(0); + throw error3; } if (asn1.derToOid(capture.contentType) !== pki2.oids.data) { - var error4 = new Error("Only PKCS#12 PFX in password integrity mode supported."); - error4.oid = asn1.derToOid(capture.contentType); - throw error4; + var error3 = new Error("Only PKCS#12 PFX in password integrity mode supported."); + error3.oid = asn1.derToOid(capture.contentType); + throw error3; } var data = capture.content.value[0]; if (data.tagClass !== asn1.Class.UNIVERSAL || data.type !== asn1.Type.OCTETSTRING) { @@ -35690,9 +35690,9 @@ var require_pkcs12 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(contentInfo, contentInfoValidator, capture, errors)) { - var error4 = new Error("Cannot read ContentInfo."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read ContentInfo."); + error3.errors = errors; + throw error3; } var obj = { encrypted: false @@ -35711,9 +35711,9 @@ var require_pkcs12 = __commonJS({ obj.encrypted = true; break; default: - var error4 = new Error("Unsupported PKCS#12 contentType."); - error4.contentType = asn1.derToOid(capture.contentType); - throw error4; + var error3 = new Error("Unsupported PKCS#12 contentType."); + error3.contentType = asn1.derToOid(capture.contentType); + throw error3; } obj.safeBags = _decodeSafeContents(safeContents, strict, password); pfx.safeContents.push(obj); @@ -35728,17 +35728,17 @@ var require_pkcs12 = __commonJS({ capture, errors )) { - var error4 = new Error("Cannot read EncryptedContentInfo."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read EncryptedContentInfo."); + error3.errors = errors; + throw error3; } var oid = asn1.derToOid(capture.contentType); if (oid !== pki2.oids.data) { - var error4 = new Error( + var error3 = new Error( "PKCS#12 EncryptedContentInfo ContentType is not Data." ); - error4.oid = oid; - throw error4; + error3.oid = oid; + throw error3; } oid = asn1.derToOid(capture.encAlgorithm); var cipher = pki2.pbe.getCipher(oid, capture.encParameter, password); @@ -35766,9 +35766,9 @@ var require_pkcs12 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(safeBag, safeBagValidator, capture, errors)) { - var error4 = new Error("Cannot read SafeBag."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read SafeBag."); + error3.errors = errors; + throw error3; } var bag = { type: asn1.derToOid(capture.bagId), @@ -35799,11 +35799,11 @@ var require_pkcs12 = __commonJS({ validator = certBagValidator; decoder = function() { if (asn1.derToOid(capture.certId) !== pki2.oids.x509Certificate) { - var error5 = new Error( + var error4 = new Error( "Unsupported certificate type, only X.509 supported." ); - error5.oid = asn1.derToOid(capture.certId); - throw error5; + error4.oid = asn1.derToOid(capture.certId); + throw error4; } var certAsn1 = asn1.fromDer(capture.cert, strict); try { @@ -35815,14 +35815,14 @@ var require_pkcs12 = __commonJS({ }; break; default: - var error4 = new Error("Unsupported PKCS#12 SafeBag type."); - error4.oid = bag.type; - throw error4; + var error3 = new Error("Unsupported PKCS#12 SafeBag type."); + error3.oid = bag.type; + throw error3; } if (validator !== void 0 && !asn1.validate(bagAsn1, validator, capture, errors)) { - var error4 = new Error("Cannot read PKCS#12 " + validator.name); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read PKCS#12 " + validator.name); + error3.errors = errors; + throw error3; } decoder(); } @@ -35835,9 +35835,9 @@ var require_pkcs12 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(attributes[i], attributeValidator, capture, errors)) { - var error4 = new Error("Cannot read PKCS#12 BagAttribute."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read PKCS#12 BagAttribute."); + error3.errors = errors; + throw error3; } var oid = asn1.derToOid(capture.oid); if (pki2.oids[oid] === void 0) { @@ -36196,9 +36196,9 @@ var require_pki = __commonJS({ pki2.privateKeyFromPem = function(pem) { var msg = forge.pem.decode(pem)[0]; if (msg.type !== "PRIVATE KEY" && msg.type !== "RSA PRIVATE KEY") { - var error4 = new Error('Could not convert private key from PEM; PEM header type is not "PRIVATE KEY" or "RSA PRIVATE KEY".'); - error4.headerType = msg.type; - throw error4; + var error3 = new Error('Could not convert private key from PEM; PEM header type is not "PRIVATE KEY" or "RSA PRIVATE KEY".'); + error3.headerType = msg.type; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { throw new Error("Could not convert private key from PEM; PEM is encrypted."); @@ -36910,7 +36910,7 @@ var require_tls = __commonJS({ }); } if (c.serverCertificate === null) { - var error4 = { + var error3 = { message: "No server certificate provided. Not enough security.", send: true, alert: { @@ -36919,21 +36919,21 @@ var require_tls = __commonJS({ } }; var depth = 0; - var ret = c.verify(c, error4.alert.description, depth, []); + var ret = c.verify(c, error3.alert.description, depth, []); if (ret !== true) { if (ret || ret === 0) { if (typeof ret === "object" && !forge.util.isArray(ret)) { if (ret.message) { - error4.message = ret.message; + error3.message = ret.message; } if (ret.alert) { - error4.alert.description = ret.alert; + error3.alert.description = ret.alert; } } else if (typeof ret === "number") { - error4.alert.description = ret; + error3.alert.description = ret; } } - return c.error(c, error4); + return c.error(c, error3); } } if (c.session.certificateRequest !== null) { @@ -37581,9 +37581,9 @@ var require_tls = __commonJS({ for (var i = 0; i < cert.length; ++i) { var msg = forge.pem.decode(cert[i])[0]; if (msg.type !== "CERTIFICATE" && msg.type !== "X509 CERTIFICATE" && msg.type !== "TRUSTED CERTIFICATE") { - var error4 = new Error('Could not convert certificate from PEM; PEM header type is not "CERTIFICATE", "X509 CERTIFICATE", or "TRUSTED CERTIFICATE".'); - error4.headerType = msg.type; - throw error4; + var error3 = new Error('Could not convert certificate from PEM; PEM header type is not "CERTIFICATE", "X509 CERTIFICATE", or "TRUSTED CERTIFICATE".'); + error3.headerType = msg.type; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { throw new Error("Could not convert certificate from PEM; PEM is encrypted."); @@ -37809,8 +37809,8 @@ var require_tls = __commonJS({ c.records = []; return c.tlsDataReady(c); }; - var _certErrorToAlertDesc = function(error4) { - switch (error4) { + var _certErrorToAlertDesc = function(error3) { + switch (error3) { case true: return true; case forge.pki.certificateError.bad_certificate: @@ -37860,19 +37860,19 @@ var require_tls = __commonJS({ var ret = c.verify(c, vfd, depth, chain2); if (ret !== true) { if (typeof ret === "object" && !forge.util.isArray(ret)) { - var error4 = new Error("The application rejected the certificate."); - error4.send = true; - error4.alert = { + var error3 = new Error("The application rejected the certificate."); + error3.send = true; + error3.alert = { level: tls.Alert.Level.fatal, description: tls.Alert.Description.bad_certificate }; if (ret.message) { - error4.message = ret.message; + error3.message = ret.message; } if (ret.alert) { - error4.alert.description = ret.alert; + error3.alert.description = ret.alert; } - throw error4; + throw error3; } if (ret !== vfd) { ret = _alertDescToCertError(ret); @@ -38956,9 +38956,9 @@ var require_ed25519 = __commonJS({ var errors = []; var valid2 = forge.asn1.validate(obj, privateKeyValidator, capture, errors); if (!valid2) { - var error4 = new Error("Invalid Key."); - error4.errors = errors; - throw error4; + var error3 = new Error("Invalid Key."); + error3.errors = errors; + throw error3; } var oid = forge.asn1.derToOid(capture.privateKeyOid); var ed25519Oid = forge.oids.EdDSA25519; @@ -38977,9 +38977,9 @@ var require_ed25519 = __commonJS({ var errors = []; var valid2 = forge.asn1.validate(obj, publicKeyValidator, capture, errors); if (!valid2) { - var error4 = new Error("Invalid Key."); - error4.errors = errors; - throw error4; + var error3 = new Error("Invalid Key."); + error3.errors = errors; + throw error3; } var oid = forge.asn1.derToOid(capture.publicKeyOid); var ed25519Oid = forge.oids.EdDSA25519; @@ -40265,9 +40265,9 @@ var require_pkcs7 = __commonJS({ p7.messageFromPem = function(pem) { var msg = forge.pem.decode(pem)[0]; if (msg.type !== "PKCS7") { - var error4 = new Error('Could not convert PKCS#7 message from PEM; PEM header type is not "PKCS#7".'); - error4.headerType = msg.type; - throw error4; + var error3 = new Error('Could not convert PKCS#7 message from PEM; PEM header type is not "PKCS#7".'); + error3.headerType = msg.type; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { throw new Error("Could not convert PKCS#7 message from PEM; PEM is encrypted."); @@ -40286,9 +40286,9 @@ var require_pkcs7 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, p7.asn1.contentInfoValidator, capture, errors)) { - var error4 = new Error("Cannot read PKCS#7 message. ASN.1 object is not an PKCS#7 ContentInfo."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read PKCS#7 message. ASN.1 object is not an PKCS#7 ContentInfo."); + error3.errors = errors; + throw error3; } var contentType = asn1.derToOid(capture.contentType); var msg; @@ -40919,9 +40919,9 @@ var require_pkcs7 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, p7.asn1.recipientInfoValidator, capture, errors)) { - var error4 = new Error("Cannot read PKCS#7 RecipientInfo. ASN.1 object is not an PKCS#7 RecipientInfo."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read PKCS#7 RecipientInfo. ASN.1 object is not an PKCS#7 RecipientInfo."); + error3.errors = errors; + throw error3; } return { version: capture.version.charCodeAt(0), @@ -41162,9 +41162,9 @@ var require_pkcs7 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, validator, capture, errors)) { - var error4 = new Error("Cannot read PKCS#7 message. ASN.1 object is not a supported PKCS#7 message."); - error4.errors = error4; - throw error4; + var error3 = new Error("Cannot read PKCS#7 message. ASN.1 object is not a supported PKCS#7 message."); + error3.errors = error3; + throw error3; } var contentType = asn1.derToOid(capture.contentType); if (contentType !== forge.pki.oids.data) { @@ -42333,8 +42333,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -43066,7 +43066,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -43075,7 +43075,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -43085,17 +43085,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -43714,7 +43714,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -43723,7 +43723,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -43733,17 +43733,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -46415,9 +46415,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -47569,8 +47569,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -47580,8 +47580,8 @@ var require_light = __commonJS({ return (await Promise.all(promises)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -47693,10 +47693,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -47730,7 +47730,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -47748,24 +47748,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } } doExpire(clearGlobalState, run, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -47775,7 +47775,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -48054,7 +48054,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve2, returned, task; + var args, cb, error3, reject, resolve2, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve2, reject } = this._queue.shift()); @@ -48065,9 +48065,9 @@ var require_light = __commonJS({ return resolve2(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -48201,8 +48201,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -48535,14 +48535,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -48838,24 +48838,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -48871,11 +48871,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -48896,12 +48896,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -55743,8 +55743,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -55978,9 +55978,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -57311,14 +57311,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -57328,7 +57328,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -57336,8 +57336,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -57557,7 +57557,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -58767,14 +58767,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -59446,11 +59446,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -59465,8 +59465,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -59960,8 +59960,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -60195,9 +60195,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -60697,8 +60697,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -60932,9 +60932,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -61997,9 +61997,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -62047,13 +62047,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -62073,21 +62073,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -62252,8 +62252,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -62659,16 +62659,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -63212,10 +63212,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -65294,12 +65294,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -65564,16 +65564,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -65698,7 +65698,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -65864,7 +65864,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -66105,9 +66105,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -66815,7 +66815,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -67878,26 +67878,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -67938,12 +67938,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -67951,13 +67951,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -67966,7 +67966,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -82572,8 +82572,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -82607,10 +82607,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -84194,8 +84194,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -84210,9 +84210,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve2, reject) => { this.emitter.on("finish", resolve2); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -85313,8 +85313,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -85398,7 +85398,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -88550,7 +88550,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -89917,9 +89917,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core12.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core12.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -90033,12 +90033,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -90072,13 +90072,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -90894,8 +90894,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -92741,8 +92741,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -94313,12 +94313,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -94338,12 +94338,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -94807,8 +94807,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -94819,8 +94819,8 @@ var require_test_transport = __commonJS({ stream.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } } @@ -95883,8 +95883,8 @@ var require_util11 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -95980,8 +95980,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -96012,18 +96012,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -96291,8 +96291,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -96493,22 +96493,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core12.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core12.error(`Failed to restore: ${error4.message}`); + core12.error(`Failed to restore: ${error3.message}`); } else { - core12.warning(`Failed to restore: ${error4.message}`); + core12.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -96563,15 +96563,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core12.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core12.error(`Failed to restore: ${error4.message}`); + core12.error(`Failed to restore: ${error3.message}`); } else { - core12.warning(`Failed to restore: ${error4.message}`); + core12.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -96579,8 +96579,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -96642,10 +96642,10 @@ var require_cache3 = __commonJS({ } core12.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core12.info(`Failed to save: ${typedError.message}`); } else { @@ -96658,8 +96658,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -96704,8 +96704,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core12.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core12.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core12.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -96724,10 +96724,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core12.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -96742,8 +96742,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -96786,14 +96786,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs.lstat(itemPath, { bigint: true }) : await fs.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs.lstat(itemPath, { bigint: true }) : await fs.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs.readdir(itemPath) : await fs.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs.readdir(itemPath) : await fs.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -96804,13 +96804,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -99469,11 +99469,11 @@ function getTestingEnvironment() { } return testingEnvironment; } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } async function checkDiskUsage(logger) { try { @@ -99496,9 +99496,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -99898,13 +99898,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core9.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -100202,9 +100202,9 @@ function isFirstPartyAnalysis(actionName) { } return process.env["CODEQL_ACTION_INIT_HAS_RUN" /* INIT_ACTION_HAS_RUN */] === "true"; } -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -100479,11 +100479,11 @@ async function runWrapper() { logger ); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); - core11.setFailed(`start-proxy action failed: ${error4.message}`); + const error3 = wrapError(unwrappedError); + core11.setFailed(`start-proxy action failed: ${error3.message}`); const errorStatusReportBase = await createStatusReportBase( "start-proxy" /* StartProxy */, - getActionsStatus(error4), + getActionsStatus(error3), startedAt, { languages: language && [language] @@ -100515,8 +100515,8 @@ async function startProxy(binPath, config, logFilePath, logger) { if (subprocess.pid) { core11.saveState("proxy-process-pid", `${subprocess.pid}`); } - subprocess.on("error", (error4) => { - subprocessError = error4; + subprocess.on("error", (error3) => { + subprocessError = error3; }); subprocess.on("exit", (code) => { if (code !== 0) { diff --git a/lib/upload-lib.js b/lib/upload-lib.js index dd1c9d395..e9e0a6747 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug4("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve6(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug4; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning8(message, properties = {}) { + exports2.error = error3; + function warning9(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning8; + exports2.warning = warning9; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -22042,8 +22042,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -22775,7 +22775,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -22784,7 +22784,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22794,17 +22794,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -23423,7 +23423,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -23432,7 +23432,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -23442,17 +23442,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -26124,9 +26124,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url2 = ""; return { value: { @@ -29208,8 +29208,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -29219,8 +29219,8 @@ var require_light = __commonJS({ return (await Promise.all(promises3)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -29332,10 +29332,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -29369,7 +29369,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -29387,24 +29387,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } } doExpire(clearGlobalState, run, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -29414,7 +29414,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -29693,7 +29693,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve6, returned, task; + var args, cb, error3, reject, resolve6, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve6, reject } = this._queue.shift()); @@ -29704,9 +29704,9 @@ var require_light = __commonJS({ return resolve6(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -29840,8 +29840,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -30174,14 +30174,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -30477,24 +30477,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -30510,11 +30510,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -30535,12 +30535,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -36085,8 +36085,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36320,9 +36320,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37653,14 +37653,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37670,7 +37670,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37678,8 +37678,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -37899,7 +37899,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39109,14 +39109,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39788,11 +39788,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39807,8 +39807,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40302,8 +40302,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40537,9 +40537,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41039,8 +41039,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41274,9 +41274,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42339,9 +42339,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42389,13 +42389,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42415,21 +42415,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42594,8 +42594,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43001,16 +43001,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43554,10 +43554,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45636,12 +45636,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -45906,16 +45906,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46040,7 +46040,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46206,7 +46206,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46447,9 +46447,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47157,7 +47157,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48220,26 +48220,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48280,12 +48280,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48293,13 +48293,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48308,7 +48308,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -62914,8 +62914,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -62949,10 +62949,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64536,8 +64536,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64552,9 +64552,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve6, reject) => { this.emitter.on("finish", resolve6); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65655,8 +65655,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65740,7 +65740,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -68892,7 +68892,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70259,9 +70259,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core12.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core12.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70375,12 +70375,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70414,13 +70414,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71236,8 +71236,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73083,8 +73083,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74655,12 +74655,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74680,12 +74680,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75149,8 +75149,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75161,8 +75161,8 @@ var require_test_transport = __commonJS({ stream2.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } } @@ -76225,8 +76225,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url2} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url2} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76322,8 +76322,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url2, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76354,18 +76354,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76633,8 +76633,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76835,22 +76835,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core12.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core12.error(`Failed to restore: ${error4.message}`); + core12.error(`Failed to restore: ${error3.message}`); } else { - core12.warning(`Failed to restore: ${error4.message}`); + core12.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76905,15 +76905,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core12.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core12.error(`Failed to restore: ${error4.message}`); + core12.error(`Failed to restore: ${error3.message}`); } else { - core12.warning(`Failed to restore: ${error4.message}`); + core12.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -76921,8 +76921,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76984,10 +76984,10 @@ var require_cache3 = __commonJS({ } core12.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core12.info(`Failed to save: ${typedError.message}`); } else { @@ -77000,8 +77000,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77046,8 +77046,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core12.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core12.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core12.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77066,10 +77066,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core12.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77084,8 +77084,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79811,7 @@ var require_debug2 = __commonJS({ if (!debug4) { try { debug4 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug4 !== "function") { debug4 = function() { @@ -79844,8 +79844,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79919,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80088,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80290,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -83276,14 +83276,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs12.lstat(itemPath, { bigint: true }) : await fs12.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs12.lstat(itemPath, { bigint: true }) : await fs12.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs12.readdir(itemPath) : await fs12.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs12.readdir(itemPath) : await fs12.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -83294,13 +83294,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -85916,9 +85916,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -86054,11 +86054,11 @@ function parseMatrixInput(matrixInput) { } return JSON.parse(matrixInput); } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } function satisfiesGHESVersion(ghesVersion, range, defaultIfInvalid) { const semverVersion = semver.coerce(ghesVersion); @@ -86492,19 +86492,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -86520,9 +86520,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -86757,13 +86757,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -90306,15 +90306,15 @@ function validateSarifFileSchema(sarif, sarifFilePath, logger) { const warnings = (result.errors ?? []).filter( (err) => err.name === "format" && typeof err.argument === "string" && warningAttributes.includes(err.argument) ); - for (const warning8 of warnings) { + for (const warning9 of warnings) { logger.info( - `Warning: '${warning8.instance}' is not a valid URI in '${warning8.property}'.` + `Warning: '${warning9.instance}' is not a valid URI in '${warning9.property}'.` ); } if (errors.length > 0) { - for (const error4 of errors) { - logger.startGroup(`Error details: ${error4.stack}`); - logger.info(JSON.stringify(error4, null, 2)); + for (const error3 of errors) { + logger.startGroup(`Error details: ${error3.stack}`); + logger.info(JSON.stringify(error3, null, 2)); logger.endGroup(); } const sarifErrors = errors.map((e) => `- ${e.stack}`); @@ -90567,10 +90567,10 @@ function shouldConsiderConfigurationError(processingErrors) { } function shouldConsiderInvalidRequest(processingErrors) { return processingErrors.every( - (error4) => error4.startsWith("rejecting SARIF") || error4.startsWith("an invalid URI was provided as a SARIF location") || error4.startsWith("locationFromSarifResult: expected artifact location") || error4.startsWith( + (error3) => error3.startsWith("rejecting SARIF") || error3.startsWith("an invalid URI was provided as a SARIF location") || error3.startsWith("locationFromSarifResult: expected artifact location") || error3.startsWith( "could not convert rules: invalid security severity value, is not a number" ) || /^SARIF URI scheme [^\s]* did not match the checkout URI scheme [^\s]*/.test( - error4 + error3 ) ); } diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index abb2273e2..96a6bc64b 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug4("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay === "number" && delay > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve2(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug4; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning9(message, properties = {}) { + exports2.error = error3; + function warning10(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning9; + exports2.warning = warning10; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -27911,8 +27911,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27922,8 @@ var require_light = __commonJS({ return (await Promise.all(promises2)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28035,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28072,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28090,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } } doExpire(clearGlobalState, run, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28117,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28396,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve2, returned, task; + var args, cb, error3, reject, resolve2, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve2, reject } = this._queue.shift()); @@ -28407,9 +28407,9 @@ var require_light = __commonJS({ return resolve2(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28543,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28877,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29180,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29213,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29238,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -31147,8 +31147,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -33460,12 +33460,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -33485,12 +33485,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -33954,8 +33954,8 @@ var require_test_transport = __commonJS({ } try { yield delay(this.responseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -33966,8 +33966,8 @@ var require_test_transport = __commonJS({ stream.notifyMessage(msg); try { yield delay(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } } @@ -36797,8 +36797,8 @@ var require_util8 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } function maskSecretUrls(body) { @@ -36891,8 +36891,8 @@ var require_artifact_twirp_client2 = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -36923,18 +36923,18 @@ var require_artifact_twirp_client2 = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -38273,8 +38273,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -38508,9 +38508,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -39841,14 +39841,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -39858,7 +39858,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -39866,8 +39866,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -40087,7 +40087,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -41297,14 +41297,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -41976,11 +41976,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -41995,8 +41995,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -42490,8 +42490,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -42725,9 +42725,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -43227,8 +43227,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -43462,9 +43462,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -44527,9 +44527,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -44577,13 +44577,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -44603,21 +44603,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -44782,8 +44782,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -45189,16 +45189,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -45742,10 +45742,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -47824,12 +47824,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -48094,16 +48094,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -48228,7 +48228,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -48394,7 +48394,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -48635,9 +48635,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -49345,7 +49345,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -50408,26 +50408,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -50468,12 +50468,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -50481,13 +50481,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -50496,7 +50496,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -65102,8 +65102,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -65137,10 +65137,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -66724,8 +66724,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -66740,9 +66740,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve2, reject) => { this.emitter.on("finish", resolve2); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -67843,8 +67843,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -67928,7 +67928,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -71080,7 +71080,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -72327,11 +72327,11 @@ var require_blob_upload = __commonJS({ blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options), chunkTimer((0, config_1.getUploadChunkTimeout)()) ]); - } catch (error4) { - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + } catch (error3) { + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } - throw error4; + throw error3; } finally { abortController.abort(); } @@ -73407,9 +73407,9 @@ var require_async = __commonJS({ invokeCallback(callback, err && (err instanceof Error || err.message) ? err : new Error(err)); }); } - function invokeCallback(callback, error4, value) { + function invokeCallback(callback, error3, value) { try { - callback(error4, value); + callback(error3, value); } catch (err) { setImmediate$1((e) => { throw e; @@ -74715,10 +74715,10 @@ var require_async = __commonJS({ function reflect(fn) { var _fn = wrapAsync(fn); return initialParams(function reflectOn(args, reflectCallback) { - args.push((error4, ...cbArgs) => { + args.push((error3, ...cbArgs) => { let retVal = {}; - if (error4) { - retVal.error = error4; + if (error3) { + retVal.error = error3; } if (cbArgs.length > 0) { var value = cbArgs; @@ -74874,13 +74874,13 @@ var require_async = __commonJS({ var timer; function timeoutCallback() { var name = asyncFn.name || "anonymous"; - var error4 = new Error('Callback function "' + name + '" timed out.'); - error4.code = "ETIMEDOUT"; + var error3 = new Error('Callback function "' + name + '" timed out.'); + error3.code = "ETIMEDOUT"; if (info7) { - error4.info = info7; + error3.info = info7; } timedOut = true; - callback(error4); + callback(error3); } args.push((...cbArgs) => { if (!timedOut) { @@ -74923,7 +74923,7 @@ var require_async = __commonJS({ return callback[PROMISE_SYMBOL]; } function tryEach(tasks, callback) { - var error4 = null; + var error3 = null; var result; return eachSeries$1(tasks, (task, taskCb) => { wrapAsync(task)((err, ...args) => { @@ -74933,10 +74933,10 @@ var require_async = __commonJS({ } else { result = args; } - error4 = err; + error3 = err; taskCb(err ? null : {}); }); - }, () => callback(error4, result)); + }, () => callback(error3, result)); } var tryEach$1 = awaitify(tryEach); function unmemoize(fn) { @@ -81416,19 +81416,19 @@ var require_from = __commonJS({ next(); } }; - readable._destroy = function(error4, cb) { + readable._destroy = function(error3, cb) { PromisePrototypeThen( - close(error4), - () => process2.nextTick(cb, error4), + close(error3), + () => process2.nextTick(cb, error3), // nextTick is here in case cb throws - (e) => process2.nextTick(cb, e || error4) + (e) => process2.nextTick(cb, e || error3) ); }; - async function close(error4) { - const hadError = error4 !== void 0 && error4 !== null; + async function close(error3) { + const hadError = error3 !== void 0 && error3 !== null; const hasThrow = typeof iterator.throw === "function"; if (hadError && hasThrow) { - const { value, done } = await iterator.throw(error4); + const { value, done } = await iterator.throw(error3); await value; if (done) { return; @@ -81636,12 +81636,12 @@ var require_readable3 = __commonJS({ this.destroy(err); }; Readable.prototype[SymbolAsyncDispose] = function() { - let error4; + let error3; if (!this.destroyed) { - error4 = this.readableEnded ? null : new AbortError(); - this.destroy(error4); + error3 = this.readableEnded ? null : new AbortError(); + this.destroy(error3); } - return new Promise2((resolve2, reject) => eos(this, (err) => err && err !== error4 ? reject(err) : resolve2(null))); + return new Promise2((resolve2, reject) => eos(this, (err) => err && err !== error3 ? reject(err) : resolve2(null))); }; Readable.prototype.push = function(chunk, encoding) { return readableAddChunk(this, chunk, encoding, false); @@ -82194,14 +82194,14 @@ var require_readable3 = __commonJS({ } } stream.on("readable", next); - let error4; + let error3; const cleanup = eos( stream, { writable: false }, (err) => { - error4 = err ? aggregateTwoErrors(error4, err) : null; + error3 = err ? aggregateTwoErrors(error3, err) : null; callback(); callback = nop; } @@ -82211,19 +82211,19 @@ var require_readable3 = __commonJS({ const chunk = stream.destroyed ? null : stream.read(); if (chunk !== null) { yield chunk; - } else if (error4) { - throw error4; - } else if (error4 === null) { + } else if (error3) { + throw error3; + } else if (error3 === null) { return; } else { await new Promise2(next); } } } catch (err) { - error4 = aggregateTwoErrors(error4, err); - throw error4; + error3 = aggregateTwoErrors(error3, err); + throw error3; } finally { - if ((error4 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error4 === void 0 || stream._readableState.autoDestroy)) { + if ((error3 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error3 === void 0 || stream._readableState.autoDestroy)) { destroyImpl.destroyer(stream, null); } else { stream.off("readable", next); @@ -83716,11 +83716,11 @@ var require_pipeline3 = __commonJS({ yield* Readable.prototype[SymbolAsyncIterator].call(val2); } async function pumpToNode(iterable, writable, finish, { end }) { - let error4; + let error3; let onresolve = null; const resume = (err) => { if (err) { - error4 = err; + error3 = err; } if (onresolve) { const callback = onresolve; @@ -83729,12 +83729,12 @@ var require_pipeline3 = __commonJS({ } }; const wait = () => new Promise2((resolve2, reject) => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { onresolve = () => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve2(); } @@ -83764,7 +83764,7 @@ var require_pipeline3 = __commonJS({ } finish(); } catch (err) { - finish(error4 !== err ? aggregateTwoErrors(error4, err) : err); + finish(error3 !== err ? aggregateTwoErrors(error3, err) : err); } finally { cleanup(); writable.off("drain", resume); @@ -83818,7 +83818,7 @@ var require_pipeline3 = __commonJS({ if (outerSignal) { disposable = addAbortListener(outerSignal, abort); } - let error4; + let error3; let value; const destroys = []; let finishCount = 0; @@ -83827,23 +83827,23 @@ var require_pipeline3 = __commonJS({ } function finishImpl(err, final) { var _disposable; - if (err && (!error4 || error4.code === "ERR_STREAM_PREMATURE_CLOSE")) { - error4 = err; + if (err && (!error3 || error3.code === "ERR_STREAM_PREMATURE_CLOSE")) { + error3 = err; } - if (!error4 && !final) { + if (!error3 && !final) { return; } while (destroys.length) { - destroys.shift()(error4); + destroys.shift()(error3); } ; (_disposable = disposable) === null || _disposable === void 0 ? void 0 : _disposable[SymbolDispose](); ac.abort(); if (final) { - if (!error4) { + if (!error3) { lastStreamCleanup.forEach((fn) => fn()); } - process2.nextTick(callback, error4, value); + process2.nextTick(callback, error3, value); } } let ret; @@ -94185,18 +94185,18 @@ var require_zip_archive_output_stream = __commonJS({ ZipArchiveOutputStream.prototype._smartStream = function(ae, callback) { var deflate = ae.getMethod() === constants.METHOD_DEFLATED; var process2 = deflate ? new DeflateCRC32Stream(this.options.zlib) : new CRC32Stream(); - var error4 = null; + var error3 = null; function handleStuff() { var digest = process2.digest().readUInt32BE(0); ae.setCrc(digest); ae.setSize(process2.size()); ae.setCompressedSize(process2.size(true)); this._afterAppend(ae); - callback(error4, ae); + callback(error3, ae); } process2.once("end", handleStuff.bind(this)); process2.once("error", function(err) { - error4 = err; + error3 = err; }); process2.pipe(this, { end: false }); return process2; @@ -95562,11 +95562,11 @@ var require_streamx = __commonJS({ } [asyncIterator]() { const stream = this; - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("readable", onreadable); this.on("close", onclose); @@ -95598,7 +95598,7 @@ var require_streamx = __commonJS({ } function ondata(data) { if (promiseReject === null) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else if (data === null && (stream._duplexState & READ_DONE) === 0) promiseReject(STREAM_DESTROYED); else promiseResolve({ value: data, done: data === null }); promiseReject = promiseResolve = null; @@ -95772,7 +95772,7 @@ var require_streamx = __commonJS({ if (all.length < 2) throw new Error("Pipeline requires at least 2 streams"); let src = all[0]; let dest = null; - let error4 = null; + let error3 = null; for (let i = 1; i < all.length; i++) { dest = all[i]; if (isStreamx(src)) { @@ -95787,14 +95787,14 @@ var require_streamx = __commonJS({ let fin = false; const autoDestroy = isStreamx(dest) || !!(dest._writableState && dest._writableState.autoDestroy); dest.on("error", (err) => { - if (error4 === null) error4 = err; + if (error3 === null) error3 = err; }); dest.on("finish", () => { fin = true; - if (!autoDestroy) done(error4); + if (!autoDestroy) done(error3); }); if (autoDestroy) { - dest.on("close", () => done(error4 || (fin ? null : PREMATURE_CLOSE))); + dest.on("close", () => done(error3 || (fin ? null : PREMATURE_CLOSE))); } } return dest; @@ -95807,8 +95807,8 @@ var require_streamx = __commonJS({ } } function onerror(err) { - if (!err || error4) return; - error4 = err; + if (!err || error3) return; + error3 = err; for (const s of all) { s.destroy(err); } @@ -96374,7 +96374,7 @@ var require_extract = __commonJS({ cb(null); } [Symbol.asyncIterator]() { - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; let entryStream = null; @@ -96382,7 +96382,7 @@ var require_extract = __commonJS({ const extract2 = this; this.on("entry", onentry); this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("close", onclose); return { @@ -96406,8 +96406,8 @@ var require_extract = __commonJS({ cb(err); } function onnext(resolve2, reject) { - if (error4) { - return reject(error4); + if (error3) { + return reject(error3); } if (entryStream) { resolve2({ value: entryStream, done: false }); @@ -96433,9 +96433,9 @@ var require_extract = __commonJS({ } } function onclose() { - consumeCallback(error4); + consumeCallback(error3); if (!promiseResolve) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else promiseResolve({ value: void 0, done: true }); promiseResolve = promiseReject = null; } @@ -97331,18 +97331,18 @@ var require_zip2 = __commonJS({ return zipUploadStream; }); } - var zipErrorCallback = (error4) => { + var zipErrorCallback = (error3) => { core14.error("An error has occurred while creating the zip file for upload"); - core14.info(error4); + core14.info(error3); throw new Error("An error has occurred during zip creation for the artifact"); }; - var zipWarningCallback = (error4) => { - if (error4.code === "ENOENT") { + var zipWarningCallback = (error3) => { + if (error3.code === "ENOENT") { core14.warning("ENOENT warning during artifact zip creation. No such file or directory"); - core14.info(error4); + core14.info(error3); } else { - core14.warning(`A non-blocking warning has occurred during artifact zip creation: ${error4.code}`); - core14.info(error4); + core14.warning(`A non-blocking warning has occurred during artifact zip creation: ${error3.code}`); + core14.info(error3); } }; var zipFinishCallback = () => { @@ -99149,8 +99149,8 @@ var require_parser_stream = __commonJS({ this.unzipStream.on("entry", function(entry) { self2.push(entry); }); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(ParserStream, Transform); @@ -99290,8 +99290,8 @@ var require_extract2 = __commonJS({ this.createdDirectories = {}; var self2 = this; this.unzipStream.on("entry", this._processEntry.bind(this)); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(Extract, Transform); @@ -99325,8 +99325,8 @@ var require_extract2 = __commonJS({ self2.unfinishedEntries--; self2._notifyAwaiter(); }); - pipedStream.on("error", function(error4) { - self2.emit("error", error4); + pipedStream.on("error", function(error3) { + self2.emit("error", error3); }); entry.pipe(pipedStream); }; @@ -99461,11 +99461,11 @@ var require_download_artifact = __commonJS({ try { yield promises_1.default.access(path2); return true; - } catch (error4) { - if (error4.code === "ENOENT") { + } catch (error3) { + if (error3.code === "ENOENT") { return false; } else { - throw error4; + throw error3; } } }); @@ -99476,9 +99476,9 @@ var require_download_artifact = __commonJS({ while (retryCount < 5) { try { return yield streamExtractExternal(url, directory); - } catch (error4) { + } catch (error3) { retryCount++; - core14.debug(`Failed to download artifact after ${retryCount} retries due to ${error4.message}. Retrying in 5 seconds...`); + core14.debug(`Failed to download artifact after ${retryCount} retries due to ${error3.message}. Retrying in 5 seconds...`); yield new Promise((resolve2) => setTimeout(resolve2, 5e3)); } } @@ -99507,10 +99507,10 @@ var require_download_artifact = __commonJS({ const extractStream = passThrough; extractStream.on("data", () => { timer.refresh(); - }).on("error", (error4) => { - core14.debug(`response.message: Artifact download failed: ${error4.message}`); + }).on("error", (error3) => { + core14.debug(`response.message: Artifact download failed: ${error3.message}`); clearTimeout(timer); - reject(error4); + reject(error3); }).pipe(unzip_stream_1.default.Extract({ path: directory })).on("close", () => { clearTimeout(timer); if (hashStream) { @@ -99519,8 +99519,8 @@ var require_download_artifact = __commonJS({ core14.info(`SHA256 digest of downloaded artifact is ${sha256Digest}`); } resolve2({ sha256Digest: `sha256:${sha256Digest}` }); - }).on("error", (error4) => { - reject(error4); + }).on("error", (error3) => { + reject(error3); }); }); }); @@ -99559,8 +99559,8 @@ var require_download_artifact = __commonJS({ core14.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -99602,8 +99602,8 @@ Are you trying to download from a different run? Try specifying a github-token w core14.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -99701,9 +99701,9 @@ var require_dist_node16 = __commonJS({ return request(options).then((response) => { octokit.log.info(`${requestOptions.method} ${path2} - ${response.status} in ${Date.now() - start}ms`); return response; - }).catch((error4) => { - octokit.log.info(`${requestOptions.method} ${path2} - ${error4.status} in ${Date.now() - start}ms`); - throw error4; + }).catch((error3) => { + octokit.log.info(`${requestOptions.method} ${path2} - ${error3.status} in ${Date.now() - start}ms`); + throw error3; }); }); } @@ -99721,22 +99721,22 @@ var require_dist_node17 = __commonJS({ return ex && typeof ex === "object" && "default" in ex ? ex["default"] : ex; } var Bottleneck = _interopDefault(require_light()); - async function errorRequest(octokit, state, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(octokit, state, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } async function wrapRequest(state, request, options) { const limiter = new Bottleneck(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -99758,12 +99758,12 @@ var require_dist_node17 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -100254,13 +100254,13 @@ var require_client2 = __commonJS({ throw new errors_1.GHESNotSupportedError(); } return (0, upload_artifact_1.uploadArtifact)(name, files, rootDirectory, options); - } catch (error4) { - (0, core_1.warning)(`Artifact upload failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Artifact upload failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions is operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -100275,13 +100275,13 @@ If the error persists, please check whether Actions is operating normally at [ht return (0, download_artifact_1.downloadArtifactPublic)(artifactId, repositoryOwner, repositoryName, token, downloadOptions); } return (0, download_artifact_1.downloadArtifactInternal)(artifactId, options); - } catch (error4) { - (0, core_1.warning)(`Download Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Download Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -100296,13 +100296,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, list_artifacts_1.listArtifactsPublic)(workflowRunId, repositoryOwner, repositoryName, token, options === null || options === void 0 ? void 0 : options.latest); } return (0, list_artifacts_1.listArtifactsInternal)(options === null || options === void 0 ? void 0 : options.latest); - } catch (error4) { - (0, core_1.warning)(`Listing Artifacts failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Listing Artifacts failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -100317,13 +100317,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, get_artifact_1.getArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, get_artifact_1.getArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Get Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Get Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -100338,13 +100338,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, delete_artifact_1.deleteArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, delete_artifact_1.deleteArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Delete Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Delete Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -100844,14 +100844,14 @@ var require_tmp = __commonJS({ options.template = _getRelativePathSync("template", options.template, tmpDir); return options; } - function _isEBADF(error4) { - return _isExpectedError(error4, -EBADF, "EBADF"); + function _isEBADF(error3) { + return _isExpectedError(error3, -EBADF, "EBADF"); } - function _isENOENT(error4) { - return _isExpectedError(error4, -ENOENT, "ENOENT"); + function _isENOENT(error3) { + return _isExpectedError(error3, -ENOENT, "ENOENT"); } - function _isExpectedError(error4, errno, code) { - return IS_WIN32 ? error4.code === code : error4.code === code && error4.errno === errno; + function _isExpectedError(error3, errno, code) { + return IS_WIN32 ? error3.code === code : error3.code === code && error3.errno === errno; } function setGracefulCleanup() { _gracefulCleanup = true; @@ -102563,9 +102563,9 @@ var require_upload_gzip = __commonJS({ const size = (yield stat(tempFilePath)).size; resolve2(size); })); - outputStream.on("error", (error4) => { - console.log(error4); - reject(error4); + outputStream.on("error", (error3) => { + console.log(error3); + reject(error3); }); }); }); @@ -102690,9 +102690,9 @@ var require_requestUtils = __commonJS({ } isRetryable = (0, utils_1.isRetryableStatusCode)(statusCode); errorMessage = `Artifact service responded with ${statusCode}`; - } catch (error4) { + } catch (error3) { isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { core14.info(`${name} - Error is not retryable`); @@ -103058,9 +103058,9 @@ var require_upload_http_client = __commonJS({ let response; try { response = yield uploadChunkRequest(); - } catch (error4) { + } catch (error3) { core14.info(`An error has been caught http-client index ${httpClientIndex}, retrying the upload`); - console.log(error4); + console.log(error3); if (incrementAndCheckRetryLimit()) { return false; } @@ -103249,8 +103249,8 @@ var require_download_http_client = __commonJS({ } this.statusReporter.incrementProcessedCount(); } - }))).catch((error4) => { - throw new Error(`Unable to download the artifact: ${error4}`); + }))).catch((error3) => { + throw new Error(`Unable to download the artifact: ${error3}`); }).finally(() => { this.statusReporter.stop(); this.downloadHttpManager.disposeAndReplaceAllClients(); @@ -103315,9 +103315,9 @@ var require_download_http_client = __commonJS({ let response; try { response = yield makeDownloadRequest(); - } catch (error4) { + } catch (error3) { core14.info("An error occurred while attempting to download a file"); - console.log(error4); + console.log(error3); yield backOff(); continue; } @@ -103331,7 +103331,7 @@ var require_download_http_client = __commonJS({ } else { forceRetry = true; } - } catch (error4) { + } catch (error3) { forceRetry = true; } } @@ -103357,31 +103357,31 @@ var require_download_http_client = __commonJS({ yield new Promise((resolve2, reject) => { if (isGzip) { const gunzip = zlib.createGunzip(); - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core14.info(`An error occurred while attempting to read the response stream`); gunzip.close(); destinationStream.close(); - reject(error4); - }).pipe(gunzip).on("error", (error4) => { + reject(error3); + }).pipe(gunzip).on("error", (error3) => { core14.info(`An error occurred while attempting to decompress the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve2(); - }).on("error", (error4) => { + }).on("error", (error3) => { core14.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } else { - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core14.info(`An error occurred while attempting to read the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve2(); - }).on("error", (error4) => { + }).on("error", (error3) => { core14.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } }); @@ -109501,9 +109501,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -109617,12 +109617,12 @@ var require_requestUtils2 = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -109656,13 +109656,13 @@ var require_requestUtils2 = __commonJS({ delay, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -110478,8 +110478,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -111241,8 +111241,8 @@ var require_util15 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -111338,8 +111338,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -111370,18 +111370,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -111649,8 +111649,8 @@ var require_tar2 = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -111851,22 +111851,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -111921,15 +111921,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -111937,8 +111937,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -112000,10 +112000,10 @@ var require_cache3 = __commonJS({ } core14.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else { @@ -112016,8 +112016,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -112062,8 +112062,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core14.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core14.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError2(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core14.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -112082,10 +112082,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -112100,8 +112100,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -114827,7 +114827,7 @@ var require_debug2 = __commonJS({ if (!debug4) { try { debug4 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug4 !== "function") { debug4 = function() { @@ -114860,8 +114860,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -114935,9 +114935,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -115104,10 +115104,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -115306,12 +115306,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -116461,14 +116461,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs2.lstat(itemPath, { bigint: true }) : await fs2.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs2.lstat(itemPath, { bigint: true }) : await fs2.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs2.readdir(itemPath) : await fs2.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs2.readdir(itemPath) : await fs2.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -116479,13 +116479,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -119170,8 +119170,8 @@ var ConfigurationError = class extends Error { super(message); } }; -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } // src/actions-util.ts @@ -119844,9 +119844,9 @@ async function runWrapper() { ) ); } - } catch (error4) { + } catch (error3) { core13.setFailed( - `upload-sarif post-action step failed: ${getErrorMessage(error4)}` + `upload-sarif post-action step failed: ${getErrorMessage(error3)}` ); } } diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 94002fa8f..6db9c67bd 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug6("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve6(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug6; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning9(message, properties = {}) { + exports2.error = error3; + function warning10(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning9; + exports2.warning = warning10; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url2 = ""; return { value: { @@ -27911,8 +27911,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27922,8 @@ var require_light = __commonJS({ return (await Promise.all(promises3)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28035,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28072,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run2, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28090,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } } doExpire(clearGlobalState, run2, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run2, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run2, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28117,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28396,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve6, returned, task; + var args, cb, error3, reject, resolve6, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve6, reject } = this._queue.shift()); @@ -28407,9 +28407,9 @@ var require_light = __commonJS({ return resolve6(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28543,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28877,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29180,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29213,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29238,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -34788,8 +34788,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -35023,9 +35023,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -36356,14 +36356,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -36373,7 +36373,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -36381,8 +36381,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -36602,7 +36602,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -37812,14 +37812,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -38491,11 +38491,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -38510,8 +38510,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -39005,8 +39005,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -39240,9 +39240,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -39742,8 +39742,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -39977,9 +39977,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -41042,9 +41042,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -41092,13 +41092,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -41118,21 +41118,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -41297,8 +41297,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -41704,16 +41704,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -42257,10 +42257,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -44339,12 +44339,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -44609,16 +44609,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -44743,7 +44743,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -44909,7 +44909,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -45150,9 +45150,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -45860,7 +45860,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -46923,26 +46923,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -46983,12 +46983,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -46996,13 +46996,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -47011,7 +47011,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -61617,8 +61617,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -61652,10 +61652,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -63239,8 +63239,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -63255,9 +63255,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve6, reject) => { this.emitter.on("finish", resolve6); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -64358,8 +64358,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -64443,7 +64443,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -67595,7 +67595,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -68962,9 +68962,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -69078,12 +69078,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -69117,13 +69117,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -69939,8 +69939,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -71786,8 +71786,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -73358,12 +73358,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -73383,12 +73383,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -73852,8 +73852,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -73864,8 +73864,8 @@ var require_test_transport = __commonJS({ stream2.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } } @@ -74928,8 +74928,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url2} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url2} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -75025,8 +75025,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url2, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -75057,18 +75057,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -75336,8 +75336,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -75538,22 +75538,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -75608,15 +75608,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -75624,8 +75624,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -75687,10 +75687,10 @@ var require_cache3 = __commonJS({ } core14.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core14.info(`Failed to save: ${typedError.message}`); } else { @@ -75703,8 +75703,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -75749,8 +75749,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core14.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core14.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core14.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -75769,10 +75769,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core14.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -75787,8 +75787,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79811,7 @@ var require_debug2 = __commonJS({ if (!debug6) { try { debug6 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug6 !== "function") { debug6 = function() { @@ -79844,8 +79844,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79919,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80088,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80290,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -83249,14 +83249,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs13.lstat(itemPath, { bigint: true }) : await fs13.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs13.lstat(itemPath, { bigint: true }) : await fs13.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs13.readdir(itemPath) : await fs13.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs13.readdir(itemPath) : await fs13.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -83267,13 +83267,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -85889,9 +85889,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -86034,11 +86034,11 @@ function parseMatrixInput(matrixInput) { } return JSON.parse(matrixInput); } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } async function checkDiskUsage(logger) { try { @@ -86061,9 +86061,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -86075,7 +86075,7 @@ function checkActionVersion(version, githubVersion) { semver.coerce(githubVersion.version) ?? "0.0.0", ">=3.20" )) { - core3.error( + core3.warning( "CodeQL Action v3 will be deprecated in December 2026. Please update all occurrences of the CodeQL Action in your workflow files to v4. For more information, see https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/" ); core3.exportVariable("CODEQL_ACTION_DID_LOG_VERSION_DEPRECATION" /* LOG_VERSION_DEPRECATION */, "true"); @@ -86540,13 +86540,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -87384,9 +87384,9 @@ function isFirstPartyAnalysis(actionName) { function isThirdPartyAnalysis(actionName) { return !isFirstPartyAnalysis(actionName); } -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -87600,19 +87600,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -87628,9 +87628,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -90806,15 +90806,15 @@ function validateSarifFileSchema(sarif, sarifFilePath, logger) { const warnings = (result.errors ?? []).filter( (err) => err.name === "format" && typeof err.argument === "string" && warningAttributes.includes(err.argument) ); - for (const warning9 of warnings) { + for (const warning10 of warnings) { logger.info( - `Warning: '${warning9.instance}' is not a valid URI in '${warning9.property}'.` + `Warning: '${warning10.instance}' is not a valid URI in '${warning10.property}'.` ); } if (errors.length > 0) { - for (const error4 of errors) { - logger.startGroup(`Error details: ${error4.stack}`); - logger.info(JSON.stringify(error4, null, 2)); + for (const error3 of errors) { + logger.startGroup(`Error details: ${error3.stack}`); + logger.info(JSON.stringify(error3, null, 2)); logger.endGroup(); } const sarifErrors = errors.map((e) => `- ${e.stack}`); @@ -91037,10 +91037,10 @@ function shouldConsiderConfigurationError(processingErrors) { } function shouldConsiderInvalidRequest(processingErrors) { return processingErrors.every( - (error4) => error4.startsWith("rejecting SARIF") || error4.startsWith("an invalid URI was provided as a SARIF location") || error4.startsWith("locationFromSarifResult: expected artifact location") || error4.startsWith( + (error3) => error3.startsWith("rejecting SARIF") || error3.startsWith("an invalid URI was provided as a SARIF location") || error3.startsWith("locationFromSarifResult: expected artifact location") || error3.startsWith( "could not convert rules: invalid security severity value, is not a number" ) || /^SARIF URI scheme [^\s]* did not match the checkout URI scheme [^\s]*/.test( - error4 + error3 ) ); } @@ -91239,18 +91239,18 @@ async function run() { logger ); } catch (unwrappedError) { - const error4 = isThirdPartyAnalysis("upload-sarif" /* UploadSarif */) && unwrappedError instanceof InvalidSarifUploadError ? new ConfigurationError(unwrappedError.message) : wrapError(unwrappedError); - const message = error4.message; + const error3 = isThirdPartyAnalysis("upload-sarif" /* UploadSarif */) && unwrappedError instanceof InvalidSarifUploadError ? new ConfigurationError(unwrappedError.message) : wrapError(unwrappedError); + const message = error3.message; core13.setFailed(message); const errorStatusReportBase = await createStatusReportBase( "upload-sarif" /* UploadSarif */, - getActionsStatus(error4), + getActionsStatus(error3), startedAt, void 0, await checkDiskUsage(logger), logger, message, - error4.stack + error3.stack ); if (errorStatusReportBase !== void 0) { await sendStatusReport(errorStatusReportBase); @@ -91261,9 +91261,9 @@ async function run() { async function runWrapper() { try { await run(); - } catch (error4) { + } catch (error3) { core13.setFailed( - `codeql/upload-sarif action failed: ${getErrorMessage(error4)}` + `codeql/upload-sarif action failed: ${getErrorMessage(error3)}` ); } } diff --git a/package-lock.json b/package-lock.json index f1a3d294f..1a2335428 100644 --- a/package-lock.json +++ b/package-lock.json @@ -1820,6 +1820,7 @@ "resolved": "https://registry.npmjs.org/@octokit/core/-/core-7.0.6.tgz", "integrity": "sha512-DhGl4xMVFGVIyMwswXeyzdL4uXD5OGILGX5N8Y+f6W7LhC1Ze2poSNrkF/fedpVDHEEZ+PHFW0vL14I+mm8K3Q==", "license": "MIT", + "peer": true, "dependencies": { "@octokit/auth-token": "^6.0.0", "@octokit/graphql": "^9.0.3", @@ -1990,6 +1991,7 @@ "resolved": "https://registry.npmjs.org/@octokit/core/-/core-5.2.2.tgz", "integrity": "sha512-/g2d4sW9nUDJOMz3mabVQvOGhVa4e/BN/Um7yca9Bb2XTzPPnfTWHWQg+IsEYO7M3Vx+EXvaM/I2pJWIMun1bg==", "license": "MIT", + "peer": true, "dependencies": { "@octokit/auth-token": "^4.0.0", "@octokit/graphql": "^7.1.0", @@ -2952,6 +2954,7 @@ "integrity": "sha512-tK3GPFWbirvNgsNKto+UmB/cRtn6TZfyw0D6IKrW55n6Vbs7KJoZtI//kpTKzE/DUmmnAFD8/Ca46s7Obs92/w==", "dev": true, "license": "MIT", + "peer": true, "dependencies": { "@typescript-eslint/scope-manager": "8.46.4", "@typescript-eslint/types": "8.46.4", @@ -3635,6 +3638,7 @@ "resolved": "https://registry.npmjs.org/acorn/-/acorn-8.15.0.tgz", "integrity": "sha512-NZyJarBfL7nWwIq+FDL6Zp/yHEhePMNnnJ0y3qfieCrmNvYct8uvtiV41UvlSe6apAfk0fY1FbWx+NwfmpvtTg==", "dev": true, + "peer": true, "bin": { "acorn": "bin/acorn" }, @@ -4291,6 +4295,7 @@ } ], "license": "MIT", + "peer": true, "dependencies": { "caniuse-lite": "^1.0.30001669", "electron-to-chromium": "^1.5.41", @@ -5144,6 +5149,7 @@ "resolved": "https://registry.npmjs.org/eslint/-/eslint-8.57.1.tgz", "integrity": "sha512-ypowyDxpVSYpkXr9WPv2PAZCtNip1Mv5KTW0SCurXv/9iOpcrH9PaqUElksqEB6pChqHGDRCFTyrZlGhnLNGiA==", "dev": true, + "peer": true, "dependencies": { "@eslint-community/eslint-utils": "^4.2.0", "@eslint-community/regexpp": "^4.6.1", @@ -5198,6 +5204,7 @@ "version": "8.3.0", "dev": true, "license": "MIT", + "peer": true, "bin": { "eslint-config-prettier": "bin/cli.js" }, @@ -5435,6 +5442,7 @@ "resolved": "https://registry.npmjs.org/eslint-plugin-import/-/eslint-plugin-import-2.29.1.tgz", "integrity": "sha512-BbPC0cuExzhiMo4Ff1BTVwHpjjv28C5R+btTOGaCRC7UEz801up0JadwkeSk5Ued6TG34uaczuVuH6qyy5YUxw==", "dev": true, + "peer": true, "dependencies": { "array-includes": "^3.1.7", "array.prototype.findlastindex": "^1.2.3", @@ -7736,6 +7744,7 @@ "resolved": "https://registry.npmjs.org/@octokit/core/-/core-7.0.6.tgz", "integrity": "sha512-DhGl4xMVFGVIyMwswXeyzdL4uXD5OGILGX5N8Y+f6W7LhC1Ze2poSNrkF/fedpVDHEEZ+PHFW0vL14I+mm8K3Q==", "license": "MIT", + "peer": true, "dependencies": { "@octokit/auth-token": "^6.0.0", "@octokit/graphql": "^9.0.3", @@ -8058,6 +8067,7 @@ "integrity": "sha512-G+YdqtITVZmOJje6QkXQWzl3fSfMxFwm1tjTyo9exhkmWSqC4Yhd1+lug++IlR2mvRVAxEDDWYkQdeSztajqgg==", "dev": true, "license": "MIT", + "peer": true, "bin": { "prettier": "bin/prettier.cjs" }, @@ -9067,6 +9077,7 @@ "integrity": "sha512-M7BAV6Rlcy5u+m6oPhAPFgJTzAioX/6B0DxyvDlo9l8+T3nLKbrczg2WLUyzd45L8RqfUMyGPzekbMvX2Ldkwg==", "dev": true, "license": "MIT", + "peer": true, "engines": { "node": ">=12" }, @@ -9284,6 +9295,7 @@ "integrity": "sha512-jl1vZzPDinLr9eUt3J/t7V6FgNEw9QjvBPdysz9KfQDD41fQrC2Y4vKQdiaUpFT4bXlb1RHhLpp8wtm6M5TgSw==", "dev": true, "license": "Apache-2.0", + "peer": true, "bin": { "tsc": "bin/tsc", "tsserver": "bin/tsserver" @@ -9357,6 +9369,7 @@ "resolved": "https://registry.npmjs.org/@typescript-eslint/parser/-/parser-8.17.0.tgz", "integrity": "sha512-Drp39TXuUlD49F7ilHHCG7TTg8IkA+hxCuULdmzWYICxGXvDXmDmWEjJYZQYgf6l/TFfYNE167m7isnc3xlIEg==", "dev": true, + "peer": true, "dependencies": { "@typescript-eslint/scope-manager": "8.17.0", "@typescript-eslint/types": "8.17.0", diff --git a/src/util.test.ts b/src/util.test.ts index 03d7d89ec..2a8d941ec 100644 --- a/src/util.test.ts +++ b/src/util.test.ts @@ -476,7 +476,7 @@ for (const [ githubVersion, )}`; test(`checkActionVersion ${reportErrorDescription} for ${versionsDescription}`, async (t) => { - const warningSpy = sinon.spy(core, "error"); + const warningSpy = sinon.spy(core, "warning"); const versionStub = sinon .stub(api, "getGitHubVersion") .resolves(githubVersion); diff --git a/src/util.ts b/src/util.ts index 7136119c5..f23c3be7d 100644 --- a/src/util.ts +++ b/src/util.ts @@ -1141,7 +1141,7 @@ export function checkActionVersion( ">=3.20", )) ) { - core.error( + core.warning( "CodeQL Action v3 will be deprecated in December 2026. " + "Please update all occurrences of the CodeQL Action in your workflow files to v4. " + "For more information, see " + From 023fd08cc92c9ae4f13726484e3dfc1c1669b0a4 Mon Sep 17 00:00:00 2001 From: Mario Campos Date: Mon, 17 Nov 2025 09:04:58 -0600 Subject: [PATCH 18/51] Add CHANGELOG.md entry for "v3 deprecation" to warning change. --- CHANGELOG.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 4ccc60273..b8c37e41e 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,7 +4,7 @@ See the [releases page](https://github.com/github/codeql-action/releases) for th ## [UNRELEASED] -No user facing changes. +- Downgraded the severity of the "v3 deprecation" notice from "error" to "warning", since the deprecation notice is meant to _warn_ users in advance. ## 4.31.3 - 13 Nov 2025 From 3b635815d6a40da3f5358c7bdffc5faf4f16ded3 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 17 Nov 2025 17:01:47 +0000 Subject: [PATCH 19/51] Bump the npm-minor group with 2 updates Bumps the npm-minor group with 2 updates: [@octokit/request-error](https://github.com/octokit/request-error.js) and [eslint-plugin-jsdoc](https://github.com/gajus/eslint-plugin-jsdoc). Updates `@octokit/request-error` from 7.0.2 to 7.1.0 - [Release notes](https://github.com/octokit/request-error.js/releases) - [Commits](https://github.com/octokit/request-error.js/compare/v7.0.2...v7.1.0) Updates `eslint-plugin-jsdoc` from 61.1.12 to 61.2.1 - [Release notes](https://github.com/gajus/eslint-plugin-jsdoc/releases) - [Changelog](https://github.com/gajus/eslint-plugin-jsdoc/blob/main/.releaserc) - [Commits](https://github.com/gajus/eslint-plugin-jsdoc/compare/v61.1.12...v61.2.1) --- updated-dependencies: - dependency-name: "@octokit/request-error" dependency-version: 7.1.0 dependency-type: direct:production update-type: version-update:semver-minor dependency-group: npm-minor - dependency-name: eslint-plugin-jsdoc dependency-version: 61.2.1 dependency-type: direct:development update-type: version-update:semver-minor dependency-group: npm-minor ... Signed-off-by: dependabot[bot] --- package-lock.json | 16 ++++++++-------- package.json | 4 ++-- 2 files changed, 10 insertions(+), 10 deletions(-) diff --git a/package-lock.json b/package-lock.json index f1a3d294f..bb8363787 100644 --- a/package-lock.json +++ b/package-lock.json @@ -20,7 +20,7 @@ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", + "@octokit/request-error": "^7.1.0", "@schemastore/package": "0.0.10", "archiver": "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -57,7 +57,7 @@ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", "glob": "^11.0.3", "nock": "^14.0.10", @@ -2315,9 +2315,9 @@ } }, "node_modules/@octokit/request-error": { - "version": "7.0.2", - "resolved": "https://registry.npmjs.org/@octokit/request-error/-/request-error-7.0.2.tgz", - "integrity": "sha512-U8piOROoQQUyExw5c6dTkU3GKxts5/ERRThIauNL7yaRoeXW0q/5bgHWT7JfWBw1UyrbK8ERId2wVkcB32n0uQ==", + "version": "7.1.0", + "resolved": "https://registry.npmjs.org/@octokit/request-error/-/request-error-7.1.0.tgz", + "integrity": "sha512-KMQIfq5sOPpkQYajXHwnhjCC0slzCNScLHs9JafXc4RAJI+9f+jNDlBNaIMTvazOPLgb4BnlhGJOTbnN0wIjPw==", "license": "MIT", "dependencies": { "@octokit/types": "^16.0.0" @@ -5470,9 +5470,9 @@ } }, "node_modules/eslint-plugin-jsdoc": { - "version": "61.1.12", - "resolved": "https://registry.npmjs.org/eslint-plugin-jsdoc/-/eslint-plugin-jsdoc-61.1.12.tgz", - "integrity": "sha512-CGJTnltz7ovwOW33xYhvA4fMuriPZpR5OnJf09SV28iU2IUpJwMd6P7zvUK8Sl56u5YzO+1F9m46wpSs2dufEw==", + "version": "61.2.1", + "resolved": "https://registry.npmjs.org/eslint-plugin-jsdoc/-/eslint-plugin-jsdoc-61.2.1.tgz", + "integrity": "sha512-Htacti3dbkNm4rlp/Bk9lqhv+gi6US9jyN22yaJ42G6wbteiTbNLChQwi25jr/BN+NOzDWhZHvCDdrhX0F8dXQ==", "dev": true, "license": "BSD-3-Clause", "dependencies": { diff --git a/package.json b/package.json index 5a3378cea..734c35a3c 100644 --- a/package.json +++ b/package.json @@ -35,7 +35,7 @@ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", + "@octokit/request-error": "^7.1.0", "@schemastore/package": "0.0.10", "archiver": "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -72,7 +72,7 @@ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", "glob": "^11.0.3", "nock": "^14.0.10", From 01577d47971e38916055ab353dc402bfcb71b7bc Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 17 Nov 2025 17:01:53 +0000 Subject: [PATCH 20/51] Bump @eslint/compat from 1.4.1 to 2.0.0 Bumps [@eslint/compat](https://github.com/eslint/rewrite/tree/HEAD/packages/compat) from 1.4.1 to 2.0.0. - [Release notes](https://github.com/eslint/rewrite/releases) - [Changelog](https://github.com/eslint/rewrite/blob/main/packages/compat/CHANGELOG.md) - [Commits](https://github.com/eslint/rewrite/commits/compat-v2.0.0/packages/compat) --- updated-dependencies: - dependency-name: "@eslint/compat" dependency-version: 2.0.0 dependency-type: direct:development update-type: version-update:semver-major ... Signed-off-by: dependabot[bot] --- package-lock.json | 54 ++++++++++++++++++++++++++++++++++++++--------- package.json | 2 +- 2 files changed, 45 insertions(+), 11 deletions(-) diff --git a/package-lock.json b/package-lock.json index f1a3d294f..b0e9855ad 100644 --- a/package-lock.json +++ b/package-lock.json @@ -36,7 +36,7 @@ }, "devDependencies": { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -1417,16 +1417,16 @@ } }, "node_modules/@eslint/compat": { - "version": "1.4.1", - "resolved": "https://registry.npmjs.org/@eslint/compat/-/compat-1.4.1.tgz", - "integrity": "sha512-cfO82V9zxxGBxcQDr1lfaYB7wykTa0b00mGa36FrJl7iTFd0Z2cHfEYuxcBRP/iNijCsWsEkA+jzT8hGYmv33w==", + "version": "2.0.0", + "resolved": "https://registry.npmjs.org/@eslint/compat/-/compat-2.0.0.tgz", + "integrity": "sha512-T9AfE1G1uv4wwq94ozgTGio5EUQBqAVe1X9qsQtSNVEYW6j3hvtZVm8Smr4qL1qDPFg+lOB2cL5RxTRMzq4CTA==", "dev": true, "license": "Apache-2.0", "dependencies": { - "@eslint/core": "^0.17.0" + "@eslint/core": "^1.0.0" }, "engines": { - "node": "^18.18.0 || ^20.9.0 || >=21.1.0" + "node": "^20.19.0 || ^22.13.0 || >=24" }, "peerDependencies": { "eslint": "^8.40 || 9" @@ -1438,16 +1438,16 @@ } }, "node_modules/@eslint/core": { - "version": "0.17.0", - "resolved": "https://registry.npmjs.org/@eslint/core/-/core-0.17.0.tgz", - "integrity": "sha512-yL/sLrpmtDaFEiUj1osRP4TI2MDz1AddJL+jZ7KSqvBuliN4xqYY54IfdN8qD8Toa6g1iloph1fxQNkjOxrrpQ==", + "version": "1.0.0", + "resolved": "https://registry.npmjs.org/@eslint/core/-/core-1.0.0.tgz", + "integrity": "sha512-PRfWP+8FOldvbApr6xL7mNCw4cJcSTq4GA7tYbgq15mRb0kWKO/wEB2jr+uwjFH3sZvEZneZyCUGTxsv4Sahyw==", "dev": true, "license": "Apache-2.0", "dependencies": { "@types/json-schema": "^7.0.15" }, "engines": { - "node": "^18.18.0 || ^20.9.0 || >=21.1.0" + "node": "^20.19.0 || ^22.13.0 || >=24" } }, "node_modules/@eslint/eslintrc": { @@ -5365,6 +5365,40 @@ "eslint": "^8 || ^9" } }, + "node_modules/eslint-plugin-github/node_modules/@eslint/compat": { + "version": "1.4.1", + "resolved": "https://registry.npmjs.org/@eslint/compat/-/compat-1.4.1.tgz", + "integrity": "sha512-cfO82V9zxxGBxcQDr1lfaYB7wykTa0b00mGa36FrJl7iTFd0Z2cHfEYuxcBRP/iNijCsWsEkA+jzT8hGYmv33w==", + "dev": true, + "license": "Apache-2.0", + "dependencies": { + "@eslint/core": "^0.17.0" + }, + "engines": { + "node": "^18.18.0 || ^20.9.0 || >=21.1.0" + }, + "peerDependencies": { + "eslint": "^8.40 || 9" + }, + "peerDependenciesMeta": { + "eslint": { + "optional": true + } + } + }, + "node_modules/eslint-plugin-github/node_modules/@eslint/core": { + "version": "0.17.0", + "resolved": "https://registry.npmjs.org/@eslint/core/-/core-0.17.0.tgz", + "integrity": "sha512-yL/sLrpmtDaFEiUj1osRP4TI2MDz1AddJL+jZ7KSqvBuliN4xqYY54IfdN8qD8Toa6g1iloph1fxQNkjOxrrpQ==", + "dev": true, + "license": "Apache-2.0", + "dependencies": { + "@types/json-schema": "^7.0.15" + }, + "engines": { + "node": "^18.18.0 || ^20.9.0 || >=21.1.0" + } + }, "node_modules/eslint-plugin-github/node_modules/debug": { "version": "3.2.7", "resolved": "https://registry.npmjs.org/debug/-/debug-3.2.7.tgz", diff --git a/package.json b/package.json index 5a3378cea..ddd27801d 100644 --- a/package.json +++ b/package.json @@ -51,7 +51,7 @@ }, "devDependencies": { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", From cd808e1260a26f39c8ccfdda9d1e7f940b0520c7 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 17 Nov 2025 17:02:13 +0000 Subject: [PATCH 21/51] Bump @types/sinon from 17.0.4 to 21.0.0 Bumps [@types/sinon](https://github.com/DefinitelyTyped/DefinitelyTyped/tree/HEAD/types/sinon) from 17.0.4 to 21.0.0. - [Release notes](https://github.com/DefinitelyTyped/DefinitelyTyped/releases) - [Commits](https://github.com/DefinitelyTyped/DefinitelyTyped/commits/HEAD/types/sinon) --- updated-dependencies: - dependency-name: "@types/sinon" dependency-version: 21.0.0 dependency-type: direct:development update-type: version-update:semver-major ... Signed-off-by: dependabot[bot] --- package-lock.json | 8 ++++---- package.json | 2 +- 2 files changed, 5 insertions(+), 5 deletions(-) diff --git a/package-lock.json b/package-lock.json index f1a3d294f..853f3e97c 100644 --- a/package-lock.json +++ b/package-lock.json @@ -47,7 +47,7 @@ "@types/node": "20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", "ava": "^6.4.1", @@ -2729,9 +2729,9 @@ "license": "MIT" }, "node_modules/@types/sinon": { - "version": "17.0.4", - "resolved": "https://registry.npmjs.org/@types/sinon/-/sinon-17.0.4.tgz", - "integrity": "sha512-RHnIrhfPO3+tJT0s7cFaXGZvsL4bbR3/k7z3P312qMS4JaS2Tk+KiwiLx1S0rQ56ERj00u1/BtdyVd0FY+Pdew==", + "version": "21.0.0", + "resolved": "https://registry.npmjs.org/@types/sinon/-/sinon-21.0.0.tgz", + "integrity": "sha512-+oHKZ0lTI+WVLxx1IbJDNmReQaIsQJjN2e7UUrJHEeByG7bFeKJYsv1E75JxTQ9QKJDp21bAa/0W2Xo4srsDnw==", "dev": true, "license": "MIT", "dependencies": { diff --git a/package.json b/package.json index 5a3378cea..7f00dfb34 100644 --- a/package.json +++ b/package.json @@ -62,7 +62,7 @@ "@types/node": "20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", "ava": "^6.4.1", From d4a7ccd1f0e0e644bdc11fda4f899cd94df27442 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 17 Nov 2025 17:03:22 +0000 Subject: [PATCH 22/51] Rebuild --- lib/analyze-action-post.js | 2 +- lib/analyze-action.js | 2 +- lib/autobuild-action.js | 2 +- lib/init-action-post.js | 2 +- lib/init-action.js | 2 +- lib/resolve-environment-action.js | 2 +- lib/setup-codeql-action.js | 2 +- lib/start-proxy-action-post.js | 2 +- lib/start-proxy-action.js | 2 +- lib/upload-lib.js | 2 +- lib/upload-sarif-action-post.js | 2 +- lib/upload-sarif-action.js | 2 +- 12 files changed, 12 insertions(+), 12 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index 7c1ecbeb0..f340b4b28 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -27678,7 +27678,7 @@ var require_package = __commonJS({ }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 2a9d16c89..1c2482ddc 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -27678,7 +27678,7 @@ var require_package = __commonJS({ }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index de693821a..39262c52e 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -27678,7 +27678,7 @@ var require_package = __commonJS({ }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", diff --git a/lib/init-action-post.js b/lib/init-action-post.js index c48e9e142..2da75956b 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -27678,7 +27678,7 @@ var require_package = __commonJS({ }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", diff --git a/lib/init-action.js b/lib/init-action.js index 2c35e5c68..e4f71da2c 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -27678,7 +27678,7 @@ var require_package = __commonJS({ }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index 4059c9db1..b556662e7 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -27678,7 +27678,7 @@ var require_package = __commonJS({ }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index 3caba058d..e7ce52832 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -27678,7 +27678,7 @@ var require_package = __commonJS({ }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index 3c7c44495..714f09214 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -27678,7 +27678,7 @@ var require_package = __commonJS({ }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index cf0fbd92d..4b56002ef 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -47336,7 +47336,7 @@ var require_package = __commonJS({ }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", diff --git a/lib/upload-lib.js b/lib/upload-lib.js index dd1c9d395..d6f066580 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -28975,7 +28975,7 @@ var require_package = __commonJS({ }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index abb2273e2..b9a43f108 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -27678,7 +27678,7 @@ var require_package = __commonJS({ }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 94002fa8f..c271c83b8 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -27678,7 +27678,7 @@ var require_package = __commonJS({ }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", From 4f39cef4c673e014515e1d5b24ccbff28dd171cd Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 17 Nov 2025 17:03:39 +0000 Subject: [PATCH 23/51] Rebuild --- lib/analyze-action-post.js | 2 +- lib/analyze-action.js | 2 +- lib/autobuild-action.js | 2 +- lib/init-action-post.js | 2 +- lib/init-action.js | 2 +- lib/resolve-environment-action.js | 2 +- lib/setup-codeql-action.js | 2 +- lib/start-proxy-action-post.js | 2 +- lib/start-proxy-action.js | 2 +- lib/upload-lib.js | 2 +- lib/upload-sarif-action-post.js | 2 +- lib/upload-sarif-action.js | 2 +- 12 files changed, 12 insertions(+), 12 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index 7c1ecbeb0..c545c844b 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -27689,7 +27689,7 @@ var require_package = __commonJS({ "@types/node": "20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 2a9d16c89..2869db7ab 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -27689,7 +27689,7 @@ var require_package = __commonJS({ "@types/node": "20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index de693821a..da1bd56d8 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -27689,7 +27689,7 @@ var require_package = __commonJS({ "@types/node": "20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", diff --git a/lib/init-action-post.js b/lib/init-action-post.js index c48e9e142..146f38776 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -27689,7 +27689,7 @@ var require_package = __commonJS({ "@types/node": "20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", diff --git a/lib/init-action.js b/lib/init-action.js index 2c35e5c68..460f1b8c4 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -27689,7 +27689,7 @@ var require_package = __commonJS({ "@types/node": "20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index 4059c9db1..1ffc88a1e 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -27689,7 +27689,7 @@ var require_package = __commonJS({ "@types/node": "20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index 3caba058d..39c3caedb 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -27689,7 +27689,7 @@ var require_package = __commonJS({ "@types/node": "20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index 3c7c44495..6ef944e1e 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -27689,7 +27689,7 @@ var require_package = __commonJS({ "@types/node": "20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index cf0fbd92d..5fb4b9116 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -47347,7 +47347,7 @@ var require_package = __commonJS({ "@types/node": "20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", diff --git a/lib/upload-lib.js b/lib/upload-lib.js index dd1c9d395..813d60f94 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -28986,7 +28986,7 @@ var require_package = __commonJS({ "@types/node": "20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index abb2273e2..00e684481 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -27689,7 +27689,7 @@ var require_package = __commonJS({ "@types/node": "20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 94002fa8f..39f11e8a0 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -27689,7 +27689,7 @@ var require_package = __commonJS({ "@types/node": "20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", From b595847fa500f410275b07bb906b143cb84f9c9d Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 17 Nov 2025 17:04:50 +0000 Subject: [PATCH 24/51] Rebuild --- lib/analyze-action-post.js | 4 ++-- lib/analyze-action.js | 4 ++-- lib/autobuild-action.js | 4 ++-- lib/init-action-post.js | 4 ++-- lib/init-action.js | 4 ++-- lib/resolve-environment-action.js | 4 ++-- lib/setup-codeql-action.js | 4 ++-- lib/start-proxy-action-post.js | 4 ++-- lib/start-proxy-action.js | 4 ++-- lib/upload-lib.js | 4 ++-- lib/upload-sarif-action-post.js | 4 ++-- lib/upload-sarif-action.js | 4 ++-- 12 files changed, 24 insertions(+), 24 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index 7c1ecbeb0..e08ccdf7f 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -27662,7 +27662,7 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", + "@octokit/request-error": "^7.1.0", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27699,7 +27699,7 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", glob: "^11.0.3", nock: "^14.0.10", diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 2a9d16c89..ab8c27f18 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -27662,7 +27662,7 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", + "@octokit/request-error": "^7.1.0", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27699,7 +27699,7 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", glob: "^11.0.3", nock: "^14.0.10", diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index de693821a..2c6b13709 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -27662,7 +27662,7 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", + "@octokit/request-error": "^7.1.0", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27699,7 +27699,7 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", glob: "^11.0.3", nock: "^14.0.10", diff --git a/lib/init-action-post.js b/lib/init-action-post.js index c48e9e142..3751f633a 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -27662,7 +27662,7 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", + "@octokit/request-error": "^7.1.0", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27699,7 +27699,7 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", glob: "^11.0.3", nock: "^14.0.10", diff --git a/lib/init-action.js b/lib/init-action.js index 2c35e5c68..a32306fb9 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -27662,7 +27662,7 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", + "@octokit/request-error": "^7.1.0", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27699,7 +27699,7 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", glob: "^11.0.3", nock: "^14.0.10", diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index 4059c9db1..4436ac21b 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -27662,7 +27662,7 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", + "@octokit/request-error": "^7.1.0", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27699,7 +27699,7 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", glob: "^11.0.3", nock: "^14.0.10", diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index 3caba058d..8a36a6ed5 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -27662,7 +27662,7 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", + "@octokit/request-error": "^7.1.0", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27699,7 +27699,7 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", glob: "^11.0.3", nock: "^14.0.10", diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index 3c7c44495..b761d478d 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -27662,7 +27662,7 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", + "@octokit/request-error": "^7.1.0", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27699,7 +27699,7 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", glob: "^11.0.3", nock: "^14.0.10", diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index cf0fbd92d..ab6349da6 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -47320,7 +47320,7 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", + "@octokit/request-error": "^7.1.0", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -47357,7 +47357,7 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", glob: "^11.0.3", nock: "^14.0.10", diff --git a/lib/upload-lib.js b/lib/upload-lib.js index dd1c9d395..bb568eda7 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -28959,7 +28959,7 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", + "@octokit/request-error": "^7.1.0", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -28996,7 +28996,7 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", glob: "^11.0.3", nock: "^14.0.10", diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index abb2273e2..e1f972a0d 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -27662,7 +27662,7 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", + "@octokit/request-error": "^7.1.0", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27699,7 +27699,7 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", glob: "^11.0.3", nock: "^14.0.10", diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 94002fa8f..4fc6e0bdf 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -27662,7 +27662,7 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", + "@octokit/request-error": "^7.1.0", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27699,7 +27699,7 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", glob: "^11.0.3", nock: "^14.0.10", From fc329e3bb50b6b174d767a52fbe4cbc4db355eaf Mon Sep 17 00:00:00 2001 From: Mario Campos Date: Mon, 17 Nov 2025 11:08:58 -0600 Subject: [PATCH 25/51] Revert "Add CHANGELOG.md entry for "v3 deprecation" to warning change." This reverts commit 023fd08cc92c9ae4f13726484e3dfc1c1669b0a4. --- CHANGELOG.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index b8c37e41e..4ccc60273 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,7 +4,7 @@ See the [releases page](https://github.com/github/codeql-action/releases) for th ## [UNRELEASED] -- Downgraded the severity of the "v3 deprecation" notice from "error" to "warning", since the deprecation notice is meant to _warn_ users in advance. +No user facing changes. ## 4.31.3 - 13 Nov 2025 From c418a0fc93ea9817b81a49f568d5714dc2bd65c6 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 17 Nov 2025 17:17:07 +0000 Subject: [PATCH 26/51] Bump ruby/setup-ruby Bumps the actions-minor group with 1 update in the /.github/workflows directory: [ruby/setup-ruby](https://github.com/ruby/setup-ruby). Updates `ruby/setup-ruby` from 1.267.0 to 1.268.0 - [Release notes](https://github.com/ruby/setup-ruby/releases) - [Changelog](https://github.com/ruby/setup-ruby/blob/master/release.rb) - [Commits](https://github.com/ruby/setup-ruby/compare/d5126b9b3579e429dd52e51e68624dda2e05be25...8aeb6ff8030dd539317f8e1769a044873b56ea71) --- updated-dependencies: - dependency-name: ruby/setup-ruby dependency-version: 1.268.0 dependency-type: direct:production update-type: version-update:semver-minor dependency-group: actions-minor ... Signed-off-by: dependabot[bot] --- .github/workflows/__rubocop-multi-language.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/__rubocop-multi-language.yml b/.github/workflows/__rubocop-multi-language.yml index 5442459ff..a5e457bb7 100644 --- a/.github/workflows/__rubocop-multi-language.yml +++ b/.github/workflows/__rubocop-multi-language.yml @@ -56,7 +56,7 @@ jobs: use-all-platform-bundle: 'false' setup-kotlin: 'true' - name: Set up Ruby - uses: ruby/setup-ruby@d5126b9b3579e429dd52e51e68624dda2e05be25 # v1.267.0 + uses: ruby/setup-ruby@8aeb6ff8030dd539317f8e1769a044873b56ea71 # v1.268.0 with: ruby-version: 2.6 - name: Install Code Scanning integration From e546fff0769babab6379aab7c5cd15f981fd3f13 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 17 Nov 2025 17:18:36 +0000 Subject: [PATCH 27/51] Rebuild --- pr-checks/checks/rubocop-multi-language.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/pr-checks/checks/rubocop-multi-language.yml b/pr-checks/checks/rubocop-multi-language.yml index 8a1f91444..192660816 100644 --- a/pr-checks/checks/rubocop-multi-language.yml +++ b/pr-checks/checks/rubocop-multi-language.yml @@ -4,7 +4,7 @@ description: "Tests using RuboCop to analyze a multi-language repository and the versions: ["default"] steps: - name: Set up Ruby - uses: ruby/setup-ruby@d5126b9b3579e429dd52e51e68624dda2e05be25 # v1.267.0 + uses: ruby/setup-ruby@8aeb6ff8030dd539317f8e1769a044873b56ea71 # v1.268.0 with: ruby-version: 2.6 - name: Install Code Scanning integration From 7bcdb4bc66db8e438dc8f9c08c766c8becf2b9c4 Mon Sep 17 00:00:00 2001 From: "Michael B. Gale" Date: Mon, 17 Nov 2025 17:48:39 +0000 Subject: [PATCH 28/51] Add additional options to PR template and clarify some --- .github/pull_request_template.md | 34 +++++++++++++++++++++++++------- 1 file changed, 27 insertions(+), 7 deletions(-) diff --git a/.github/pull_request_template.md b/.github/pull_request_template.md index 3b632d8f3..3c6d14f71 100644 --- a/.github/pull_request_template.md +++ b/.github/pull_request_template.md @@ -18,14 +18,25 @@ For internal use only. Please select the risk level of this change: #### Which use cases does this change impact? - + -- **Advanced setup** - Impacts users who have custom workflows. -- **Default setup** - Impacts users who use default setup. -- **Code Scanning** - Impacts Code Scanning (i.e. `analysis-kinds: code-scanning`). -- **Code Quality** - Impacts Code Quality (i.e. `analysis-kinds: code-quality`). -- **Third-party analyses** - Impacts third-party analyses (i.e. `upload-sarif`). -- **GHES** - Impacts GitHub Enterprise Server. +Workflow types: + +- **Advanced setup** - Impacts users who have custom CodeQL workflows. +- **Managed** - Impacts users with `dynamic` workflows (Default Setup, CCR, ...). + +Products: + +- **Code Scanning** - The changes impact analyses when `analysis-kinds: code-scanning`. +- **Code Quality** - The changes impact analyses when `analysis-kinds: code-quality`. +- **CCR** - The changes impact analyses for Copilot Code Reviews. +- **Third-party analyses** - The changes affect the `upload-sarif` action. + +Environments: + +- **Dotcom** - Impacts CodeQL workflows on `github.com`. +- **GHES** - Impacts CodeQL workflows on GitHub Enterprise Server. +- **Testing/None** - This change does not impact any CodeQL workflows in production. #### How did/will you validate this change? @@ -54,6 +65,15 @@ For internal use only. Please select the risk level of this change: - **Alerts** - New or existing monitors will trip if something goes wrong with this change. - **Other** - Please provide details. +#### Are there any special considerations for merging or releasing this change? + + + +- **No special considerations** - This change can be merged at any time. +- **Special considerations** - This change should only be merged once certain preconditions are met. Please provide details of those or link to this PR from an internal issue. + ### Merge / deployment checklist - Confirm this change is backwards compatible with existing workflows. From 4f746e4a60b95e41fe6c91e3f14806cd10a16621 Mon Sep 17 00:00:00 2001 From: Kasper Svendsen Date: Tue, 18 Nov 2025 08:11:04 +0100 Subject: [PATCH 29/51] Overlay: Fall back to full analysis if runner disk space is low --- lib/analyze-action-post.js | 2 + lib/analyze-action.js | 2 + lib/autobuild-action.js | 2 + lib/init-action-post.js | 2 + lib/init-action.js | 44 ++++++++----- lib/resolve-environment-action.js | 2 + lib/setup-codeql-action.js | 2 + lib/start-proxy-action-post.js | 2 + lib/start-proxy-action.js | 2 + lib/upload-lib.js | 2 + lib/upload-sarif-action-post.js | 2 + lib/upload-sarif-action.js | 2 + src/config-utils.test.ts | 100 ++++++++++++++++++++++++++++++ src/config-utils.ts | 65 +++++++++++++------ 14 files changed, 196 insertions(+), 35 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index d815da082..3dcda989f 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -120102,6 +120102,8 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 5dcfd5062..11b32eac9 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -89370,6 +89370,8 @@ async function cachePrefix(codeql, language) { } // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index 2655e4315..0a5626575 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -84416,6 +84416,8 @@ var GitHubFeatureFlags = class { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, diff --git a/lib/init-action-post.js b/lib/init-action-post.js index c238f7d06..6ab23eb5a 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -123766,6 +123766,8 @@ ${jsonContents}` var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, diff --git a/lib/init-action.js b/lib/init-action.js index 6de0e4e2d..0630127df 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -86653,6 +86653,8 @@ async function cachePrefix(codeql, language) { } // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; async function getSupportedLanguageMap(codeql, logger) { const resolveSupportedLanguagesUsingCli = await codeql.supportsFeature( "builtinExtractorsSpecifyDefaultQueries" /* BuiltinExtractorsSpecifyDefaultQueries */ @@ -86918,24 +86920,34 @@ async function getOverlayDatabaseMode(codeql, features, languages, sourceRoot, b logger.info( `Setting overlay database mode to ${overlayDatabaseMode} from the CODEQL_OVERLAY_DATABASE_MODE environment variable.` ); - } else if (await isOverlayAnalysisFeatureEnabled( - features, - codeql, - languages, - codeScanningConfig - )) { - if (isAnalyzingPullRequest()) { - overlayDatabaseMode = "overlay" /* Overlay */; - useOverlayDatabaseCaching = true; + } else { + const diskUsage = await checkDiskUsage(logger); + if (diskUsage === void 0 || diskUsage.numAvailableBytes < OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES) { + const diskSpaceMb = diskUsage === void 0 ? 0 : diskUsage.numAvailableBytes / 1e6; + overlayDatabaseMode = "none" /* None */; + useOverlayDatabaseCaching = false; logger.info( - `Setting overlay database mode to ${overlayDatabaseMode} with caching because we are analyzing a pull request.` - ); - } else if (await isAnalyzingDefaultBranch()) { - overlayDatabaseMode = "overlay-base" /* OverlayBase */; - useOverlayDatabaseCaching = true; - logger.info( - `Setting overlay database mode to ${overlayDatabaseMode} with caching because we are analyzing the default branch.` + `Setting overlay database mode to ${overlayDatabaseMode} due to insufficient disk space (${diskSpaceMb} MB).` ); + } else if (await isOverlayAnalysisFeatureEnabled( + features, + codeql, + languages, + codeScanningConfig + )) { + if (isAnalyzingPullRequest()) { + overlayDatabaseMode = "overlay" /* Overlay */; + useOverlayDatabaseCaching = true; + logger.info( + `Setting overlay database mode to ${overlayDatabaseMode} with caching because we are analyzing a pull request.` + ); + } else if (await isAnalyzingDefaultBranch()) { + overlayDatabaseMode = "overlay-base" /* OverlayBase */; + useOverlayDatabaseCaching = true; + logger.info( + `Setting overlay database mode to ${overlayDatabaseMode} with caching because we are analyzing the default branch.` + ); + } } } const nonOverlayAnalysis = { diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index 8a757d8c6..3cca7431d 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -84142,6 +84142,8 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index 31f135f53..0d0df74a4 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -84587,6 +84587,8 @@ var PACK_IDENTIFIER_PATTERN = (function() { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index 86991078a..9377bcea4 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -119508,6 +119508,8 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index 3b6d4deca..26d0fa13b 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -100170,6 +100170,8 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, diff --git a/lib/upload-lib.js b/lib/upload-lib.js index e9e0a6747..be50f109a 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -87226,6 +87226,8 @@ ${jsonContents}` var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index 96a6bc64b..af2299f7f 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -119674,6 +119674,8 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 6db9c67bd..ea4b0e31b 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -87307,6 +87307,8 @@ ${jsonContents}` var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, diff --git a/src/config-utils.test.ts b/src/config-utils.test.ts index da58dd8b1..b9407152b 100644 --- a/src/config-utils.test.ts +++ b/src/config-utils.test.ts @@ -37,7 +37,9 @@ import { ConfigurationError, withTmpDir, BuildMode, + DiskUsage, } from "./util"; +import * as util from "./util"; setupTests(test); @@ -995,6 +997,7 @@ interface OverlayDatabaseModeTestSetup { codeqlVersion: string; gitRoot: string | undefined; codeScanningConfig: configUtils.UserConfig; + diskUsage: DiskUsage | undefined; } const defaultOverlayDatabaseModeTestSetup: OverlayDatabaseModeTestSetup = { @@ -1007,6 +1010,10 @@ const defaultOverlayDatabaseModeTestSetup: OverlayDatabaseModeTestSetup = { codeqlVersion: CODEQL_OVERLAY_MINIMUM_VERSION, gitRoot: "/some/git/root", codeScanningConfig: {}, + diskUsage: { + numAvailableBytes: 50_000_000_000, + numTotalBytes: 100_000_000_000, + }, }; const getOverlayDatabaseModeMacro = test.macro({ @@ -1039,6 +1046,8 @@ const getOverlayDatabaseModeMacro = test.macro({ setup.overlayDatabaseEnvVar; } + sinon.stub(util, "checkDiskUsage").resolves(setup.diskUsage); + // Mock feature flags const features = createFeatures(setup.features); @@ -1196,6 +1205,45 @@ test( }, ); +test( + getOverlayDatabaseModeMacro, + "No overlay-base database on default branch if runner disk space is too low", + { + languages: [KnownLanguage.javascript], + features: [ + Feature.OverlayAnalysis, + Feature.OverlayAnalysisCodeScanningJavascript, + ], + isDefaultBranch: true, + diskUsage: { + numAvailableBytes: 1_000_000_000, + numTotalBytes: 100_000_000_000, + }, + }, + { + overlayDatabaseMode: OverlayDatabaseMode.None, + useOverlayDatabaseCaching: false, + }, +); + +test( + getOverlayDatabaseModeMacro, + "No overlay-base database on default branch if we can't determine runner disk space", + { + languages: [KnownLanguage.javascript], + features: [ + Feature.OverlayAnalysis, + Feature.OverlayAnalysisCodeScanningJavascript, + ], + isDefaultBranch: true, + diskUsage: undefined, + }, + { + overlayDatabaseMode: OverlayDatabaseMode.None, + useOverlayDatabaseCaching: false, + }, +); + test( getOverlayDatabaseModeMacro, "No overlay-base database on default branch when code-scanning feature enabled with disable-default-queries", @@ -1366,6 +1414,45 @@ test( }, ); +test( + getOverlayDatabaseModeMacro, + "No overlay analysis on PR if runner disk space is too low", + { + languages: [KnownLanguage.javascript], + features: [ + Feature.OverlayAnalysis, + Feature.OverlayAnalysisCodeScanningJavascript, + ], + isPullRequest: true, + diskUsage: { + numAvailableBytes: 1_000_000_000, + numTotalBytes: 100_000_000_000, + }, + }, + { + overlayDatabaseMode: OverlayDatabaseMode.None, + useOverlayDatabaseCaching: false, + }, +); + +test( + getOverlayDatabaseModeMacro, + "No overlay analysis on PR if we can't determine runner disk space", + { + languages: [KnownLanguage.javascript], + features: [ + Feature.OverlayAnalysis, + Feature.OverlayAnalysisCodeScanningJavascript, + ], + isPullRequest: true, + diskUsage: undefined, + }, + { + overlayDatabaseMode: OverlayDatabaseMode.None, + useOverlayDatabaseCaching: false, + }, +); + test( getOverlayDatabaseModeMacro, "No overlay analysis on PR when code-scanning feature enabled with disable-default-queries", @@ -1500,6 +1587,19 @@ test( }, ); +test( + getOverlayDatabaseModeMacro, + "Overlay PR analysis by env on a runner with low disk space", + { + overlayDatabaseEnvVar: "overlay", + diskUsage: { numAvailableBytes: 0, numTotalBytes: 100_000_000_000 }, + }, + { + overlayDatabaseMode: OverlayDatabaseMode.Overlay, + useOverlayDatabaseCaching: false, + }, +); + test( getOverlayDatabaseModeMacro, "Overlay PR analysis by feature flag", diff --git a/src/config-utils.ts b/src/config-utils.ts index fa88d4b4a..6a1c40b13 100644 --- a/src/config-utils.ts +++ b/src/config-utils.ts @@ -43,10 +43,22 @@ import { codeQlVersionAtLeast, cloneObject, isDefined, + checkDiskUsage, } from "./util"; export * from "./config/db-config"; +/** + * The minimum available disk space (in MB) required to perform overlay analysis. + * If the available disk space on the runner is below the threshold when deciding + * whether to perform overlay analysis, then the action will not perform overlay + * analysis unless overlay analysis has been explicitly enabled via environment + * variable. + */ +const OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15000; +const OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = + OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1_000_000; + export type RegistryConfigWithCredentials = RegistryConfigNoCredentials & { // Token to use when downloading packs from this registry. token: string; @@ -667,28 +679,43 @@ export async function getOverlayDatabaseMode( `Setting overlay database mode to ${overlayDatabaseMode} ` + "from the CODEQL_OVERLAY_DATABASE_MODE environment variable.", ); - } else if ( - await isOverlayAnalysisFeatureEnabled( - features, - codeql, - languages, - codeScanningConfig, - ) - ) { - if (isAnalyzingPullRequest()) { - overlayDatabaseMode = OverlayDatabaseMode.Overlay; - useOverlayDatabaseCaching = true; + } else { + const diskUsage = await checkDiskUsage(logger); + if ( + diskUsage === undefined || + diskUsage.numAvailableBytes < OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES + ) { + const diskSpaceMb = + diskUsage === undefined ? 0 : diskUsage.numAvailableBytes / 1_000_000; + overlayDatabaseMode = OverlayDatabaseMode.None; + useOverlayDatabaseCaching = false; logger.info( `Setting overlay database mode to ${overlayDatabaseMode} ` + - "with caching because we are analyzing a pull request.", - ); - } else if (await isAnalyzingDefaultBranch()) { - overlayDatabaseMode = OverlayDatabaseMode.OverlayBase; - useOverlayDatabaseCaching = true; - logger.info( - `Setting overlay database mode to ${overlayDatabaseMode} ` + - "with caching because we are analyzing the default branch.", + `due to insufficient disk space (${diskSpaceMb} MB).`, ); + } else if ( + await isOverlayAnalysisFeatureEnabled( + features, + codeql, + languages, + codeScanningConfig, + ) + ) { + if (isAnalyzingPullRequest()) { + overlayDatabaseMode = OverlayDatabaseMode.Overlay; + useOverlayDatabaseCaching = true; + logger.info( + `Setting overlay database mode to ${overlayDatabaseMode} ` + + "with caching because we are analyzing a pull request.", + ); + } else if (await isAnalyzingDefaultBranch()) { + overlayDatabaseMode = OverlayDatabaseMode.OverlayBase; + useOverlayDatabaseCaching = true; + logger.info( + `Setting overlay database mode to ${overlayDatabaseMode} ` + + "with caching because we are analyzing the default branch.", + ); + } } } From 528362a7c177806bfb952333f21e18a1721bed2f Mon Sep 17 00:00:00 2001 From: "Michael B. Gale" Date: Tue, 18 Nov 2025 10:49:13 +0000 Subject: [PATCH 30/51] Bump `glob` to at least `11.1.0` --- lib/analyze-action-post.js | 370 ++++++++++++++++++----------- lib/analyze-action.js | 7 +- lib/autobuild-action.js | 7 +- lib/init-action-post.js | 370 ++++++++++++++++++----------- lib/init-action.js | 7 +- lib/resolve-environment-action.js | 7 +- lib/setup-codeql-action.js | 7 +- lib/start-proxy-action-post.js | 370 ++++++++++++++++++----------- lib/start-proxy-action.js | 7 +- lib/upload-lib.js | 7 +- lib/upload-sarif-action-post.js | 374 +++++++++++++++++++----------- lib/upload-sarif-action.js | 7 +- package-lock.json | 252 +------------------- package.json | 7 +- 14 files changed, 1008 insertions(+), 791 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index d815da082..7dde941ee 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -27694,14 +27694,14 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", - eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + eslint: "^8.57.1", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27725,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -97357,52 +97358,128 @@ var require_isPlainObject = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/brace-expansion/index.js -var require_brace_expansion3 = __commonJS({ - "node_modules/archiver-utils/node_modules/brace-expansion/index.js"(exports2, module2) { - var balanced = require_balanced_match(); - module2.exports = expandTop; +// node_modules/@isaacs/balanced-match/dist/commonjs/index.js +var require_commonjs13 = __commonJS({ + "node_modules/@isaacs/balanced-match/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.range = exports2.balanced = void 0; + var balanced = (a, b, str2) => { + const ma = a instanceof RegExp ? maybeMatch(a, str2) : a; + const mb = b instanceof RegExp ? maybeMatch(b, str2) : b; + const r = ma !== null && mb != null && (0, exports2.range)(ma, mb, str2); + return r && { + start: r[0], + end: r[1], + pre: str2.slice(0, r[0]), + body: str2.slice(r[0] + ma.length, r[1]), + post: str2.slice(r[1] + mb.length) + }; + }; + exports2.balanced = balanced; + var maybeMatch = (reg, str2) => { + const m = str2.match(reg); + return m ? m[0] : null; + }; + var range = (a, b, str2) => { + let begs, beg, left, right = void 0, result; + let ai = str2.indexOf(a); + let bi = str2.indexOf(b, ai + 1); + let i = ai; + if (ai >= 0 && bi > 0) { + if (a === b) { + return [ai, bi]; + } + begs = []; + left = str2.length; + while (i >= 0 && !result) { + if (i === ai) { + begs.push(i); + ai = str2.indexOf(a, i + 1); + } else if (begs.length === 1) { + const r = begs.pop(); + if (r !== void 0) + result = [r, bi]; + } else { + beg = begs.pop(); + if (beg !== void 0 && beg < left) { + left = beg; + right = bi; + } + bi = str2.indexOf(b, i + 1); + } + i = ai < bi && ai >= 0 ? ai : bi; + } + if (begs.length && right !== void 0) { + result = [left, right]; + } + } + return result; + }; + exports2.range = range; + } +}); + +// node_modules/@isaacs/brace-expansion/dist/commonjs/index.js +var require_commonjs14 = __commonJS({ + "node_modules/@isaacs/brace-expansion/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.expand = expand; + var balanced_match_1 = require_commonjs13(); var escSlash = "\0SLASH" + Math.random() + "\0"; var escOpen = "\0OPEN" + Math.random() + "\0"; var escClose = "\0CLOSE" + Math.random() + "\0"; var escComma = "\0COMMA" + Math.random() + "\0"; var escPeriod = "\0PERIOD" + Math.random() + "\0"; + var escSlashPattern = new RegExp(escSlash, "g"); + var escOpenPattern = new RegExp(escOpen, "g"); + var escClosePattern = new RegExp(escClose, "g"); + var escCommaPattern = new RegExp(escComma, "g"); + var escPeriodPattern = new RegExp(escPeriod, "g"); + var slashPattern = /\\\\/g; + var openPattern = /\\{/g; + var closePattern = /\\}/g; + var commaPattern = /\\,/g; + var periodPattern = /\\./g; function numeric(str2) { - return parseInt(str2, 10) == str2 ? parseInt(str2, 10) : str2.charCodeAt(0); + return !isNaN(str2) ? parseInt(str2, 10) : str2.charCodeAt(0); } function escapeBraces(str2) { - return str2.split("\\\\").join(escSlash).split("\\{").join(escOpen).split("\\}").join(escClose).split("\\,").join(escComma).split("\\.").join(escPeriod); + return str2.replace(slashPattern, escSlash).replace(openPattern, escOpen).replace(closePattern, escClose).replace(commaPattern, escComma).replace(periodPattern, escPeriod); } function unescapeBraces(str2) { - return str2.split(escSlash).join("\\").split(escOpen).join("{").split(escClose).join("}").split(escComma).join(",").split(escPeriod).join("."); + return str2.replace(escSlashPattern, "\\").replace(escOpenPattern, "{").replace(escClosePattern, "}").replace(escCommaPattern, ",").replace(escPeriodPattern, "."); } function parseCommaParts(str2) { - if (!str2) + if (!str2) { return [""]; - var parts = []; - var m = balanced("{", "}", str2); - if (!m) + } + const parts = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) { return str2.split(","); - var pre = m.pre; - var body = m.body; - var post = m.post; - var p = pre.split(","); + } + const { pre, body, post } = m; + const p = pre.split(","); p[p.length - 1] += "{" + body + "}"; - var postParts = parseCommaParts(post); + const postParts = parseCommaParts(post); if (post.length) { + ; p[p.length - 1] += postParts.shift(); p.push.apply(p, postParts); } parts.push.apply(parts, p); return parts; } - function expandTop(str2) { - if (!str2) + function expand(str2) { + if (!str2) { return []; - if (str2.substr(0, 2) === "{}") { - str2 = "\\{\\}" + str2.substr(2); } - return expand(escapeBraces(str2), true).map(unescapeBraces); + if (str2.slice(0, 2) === "{}") { + str2 = "\\{\\}" + str2.slice(2); + } + return expand_(escapeBraces(str2), true).map(unescapeBraces); } function embrace(str2) { return "{" + str2 + "}"; @@ -97416,73 +97493,74 @@ var require_brace_expansion3 = __commonJS({ function gte5(i, y) { return i >= y; } - function expand(str2, isTop) { - var expansions = []; - var m = balanced("{", "}", str2); - if (!m) return [str2]; - var pre = m.pre; - var post = m.post.length ? expand(m.post, false) : [""]; + function expand_(str2, isTop) { + const expansions = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) + return [str2]; + const pre = m.pre; + const post = m.post.length ? expand_(m.post, false) : [""]; if (/\$$/.test(m.pre)) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + "{" + m.body + "}" + post[k]; + for (let k = 0; k < post.length; k++) { + const expansion = pre + "{" + m.body + "}" + post[k]; expansions.push(expansion); } } else { - var isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); - var isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); - var isSequence = isNumericSequence || isAlphaSequence; - var isOptions = m.body.indexOf(",") >= 0; + const isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); + const isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); + const isSequence = isNumericSequence || isAlphaSequence; + const isOptions = m.body.indexOf(",") >= 0; if (!isSequence && !isOptions) { if (m.post.match(/,(?!,).*\}/)) { str2 = m.pre + "{" + m.body + escClose + m.post; - return expand(str2); + return expand_(str2); } return [str2]; } - var n; + let n; if (isSequence) { n = m.body.split(/\.\./); } else { n = parseCommaParts(m.body); - if (n.length === 1) { - n = expand(n[0], false).map(embrace); + if (n.length === 1 && n[0] !== void 0) { + n = expand_(n[0], false).map(embrace); if (n.length === 1) { - return post.map(function(p) { - return m.pre + n[0] + p; - }); + return post.map((p) => m.pre + n[0] + p); } } } - var N; - if (isSequence) { - var x = numeric(n[0]); - var y = numeric(n[1]); - var width = Math.max(n[0].length, n[1].length); - var incr = n.length == 3 ? Math.abs(numeric(n[2])) : 1; - var test = lte; - var reverse = y < x; + let N; + if (isSequence && n[0] !== void 0 && n[1] !== void 0) { + const x = numeric(n[0]); + const y = numeric(n[1]); + const width = Math.max(n[0].length, n[1].length); + let incr = n.length === 3 && n[2] !== void 0 ? Math.abs(numeric(n[2])) : 1; + let test = lte; + const reverse = y < x; if (reverse) { incr *= -1; test = gte5; } - var pad = n.some(isPadded); + const pad = n.some(isPadded); N = []; - for (var i = x; test(i, y); i += incr) { - var c; + for (let i = x; test(i, y); i += incr) { + let c; if (isAlphaSequence) { c = String.fromCharCode(i); - if (c === "\\") + if (c === "\\") { c = ""; + } } else { c = String(i); if (pad) { - var need = width - c.length; + const need = width - c.length; if (need > 0) { - var z = new Array(need + 1).join("0"); - if (i < 0) + const z = new Array(need + 1).join("0"); + if (i < 0) { c = "-" + z + c.slice(1); - else + } else { c = z + c; + } } } } @@ -97490,15 +97568,16 @@ var require_brace_expansion3 = __commonJS({ } } else { N = []; - for (var j = 0; j < n.length; j++) { - N.push.apply(N, expand(n[j], false)); + for (let j = 0; j < n.length; j++) { + N.push.apply(N, expand_(n[j], false)); } } - for (var j = 0; j < N.length; j++) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + N[j] + post[k]; - if (!isTop || isSequence || expansion) + for (let j = 0; j < N.length; j++) { + for (let k = 0; k < post.length; k++) { + const expansion = pre + N[j] + post[k]; + if (!isTop || isSequence || expansion) { expansions.push(expansion); + } } } } @@ -97507,9 +97586,9 @@ var require_brace_expansion3 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js var require_assert_valid_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.assertValidPattern = void 0; @@ -97526,9 +97605,9 @@ var require_assert_valid_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js var require_brace_expressions = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.parseClass = void 0; @@ -97643,22 +97722,25 @@ var require_brace_expressions = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js var require_unescape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = void 0; - var unescape = (s, { windowsPathsNoEscape = false } = {}) => { - return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + var unescape = (s, { windowsPathsNoEscape = false, magicalBraces = true } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + } + return windowsPathsNoEscape ? s.replace(/\[([^\/\\{}])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\{}])\]/g, "$1$2").replace(/\\([^\/{}])/g, "$1"); }; exports2.unescape = unescape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js var require_ast = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.AST = void 0; @@ -98014,7 +98096,7 @@ var require_ast = __commonJS({ if (this.#root === this) this.#fillNegs(); if (!this.type) { - const noEmpty = this.isStart() && this.isEnd(); + const noEmpty = this.isStart() && this.isEnd() && !this.#parts.some((s) => typeof s !== "string"); const src = this.#parts.map((p) => { const [re, _2, hasMagic, uflag] = typeof p === "string" ? _AST.#parseGlob(p, this.#hasMagic, noEmpty) : p.toRegExpSource(allowDot); this.#hasMagic = this.#hasMagic || hasMagic; @@ -98124,10 +98206,7 @@ var require_ast = __commonJS({ } } if (c === "*") { - if (noEmpty && glob2 === "*") - re += starNoEmpty; - else - re += star; + re += noEmpty && glob2 === "*" ? starNoEmpty : star; hasMagic = true; continue; } @@ -98145,29 +98224,29 @@ var require_ast = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js var require_escape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.escape = void 0; - var escape = (s, { windowsPathsNoEscape = false } = {}) => { + var escape = (s, { windowsPathsNoEscape = false, magicalBraces = false } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/[?*()[\]{}]/g, "[$&]") : s.replace(/[?*()[\]\\{}]/g, "\\$&"); + } return windowsPathsNoEscape ? s.replace(/[?*()[\]]/g, "[$&]") : s.replace(/[?*()[\]\\]/g, "\\$&"); }; exports2.escape = escape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js -var require_commonjs13 = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js"(exports2) { +// node_modules/glob/node_modules/minimatch/dist/commonjs/index.js +var require_commonjs15 = __commonJS({ + "node_modules/glob/node_modules/minimatch/dist/commonjs/index.js"(exports2) { "use strict"; - var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { - return mod && mod.__esModule ? mod : { "default": mod }; - }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = exports2.escape = exports2.AST = exports2.Minimatch = exports2.match = exports2.makeRe = exports2.braceExpand = exports2.defaults = exports2.filter = exports2.GLOBSTAR = exports2.sep = exports2.minimatch = void 0; - var brace_expansion_1 = __importDefault4(require_brace_expansion3()); + var brace_expansion_1 = require_commonjs14(); var assert_valid_pattern_js_1 = require_assert_valid_pattern(); var ast_js_1 = require_ast(); var escape_js_1 = require_escape(); @@ -98290,7 +98369,7 @@ var require_commonjs13 = __commonJS({ if (options.nobrace || !/\{(?:(?!\{).)*\}/.test(pattern)) { return [pattern]; } - return (0, brace_expansion_1.default)(pattern); + return (0, brace_expansion_1.expand)(pattern); }; exports2.braceExpand = braceExpand; exports2.minimatch.braceExpand = exports2.braceExpand; @@ -98814,16 +98893,27 @@ var require_commonjs13 = __commonJS({ pp[i] = twoStar; } } else if (next === void 0) { - pp[i - 1] = prev + "(?:\\/|" + twoStar + ")?"; + pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + ")?"; } else if (next !== exports2.GLOBSTAR) { pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + "\\/)" + next; pp[i + 1] = exports2.GLOBSTAR; } }); - return pp.filter((p) => p !== exports2.GLOBSTAR).join("/"); + const filtered = pp.filter((p) => p !== exports2.GLOBSTAR); + if (this.partial && filtered.length >= 1) { + const prefixes = []; + for (let i = 1; i <= filtered.length; i++) { + prefixes.push(filtered.slice(0, i).join("/")); + } + return "(?:" + prefixes.join("|") + ")"; + } + return filtered.join("/"); }).join("|"); const [open, close] = set2.length > 1 ? ["(?:", ")"] : ["", ""]; re = "^" + open + re + close + "$"; + if (this.partial) { + re = "^(?:\\/|" + open + re.slice(1, -1) + close + ")$"; + } if (this.negate) re = "^(?!" + re + ").+$"; try { @@ -98910,9 +99000,9 @@ var require_commonjs13 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js -var require_commonjs14 = __commonJS({ - "node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js"(exports2) { +// node_modules/lru-cache/dist/commonjs/index.js +var require_commonjs16 = __commonJS({ + "node_modules/lru-cache/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.LRUCache = void 0; @@ -99001,6 +99091,7 @@ var require_commonjs14 = __commonJS({ #max; #maxSize; #dispose; + #onInsert; #disposeAfter; #fetchMethod; #memoMethod; @@ -99082,6 +99173,7 @@ var require_commonjs14 = __commonJS({ #hasDispose; #hasFetchMethod; #hasDisposeAfter; + #hasOnInsert; /** * Do not call this method unless you need to inspect the * inner workings of the cache. If anything returned by this @@ -99158,6 +99250,12 @@ var require_commonjs14 = __commonJS({ get dispose() { return this.#dispose; } + /** + * {@link LRUCache.OptionsBase.onInsert} (read-only) + */ + get onInsert() { + return this.#onInsert; + } /** * {@link LRUCache.OptionsBase.disposeAfter} (read-only) */ @@ -99165,7 +99263,7 @@ var require_commonjs14 = __commonJS({ return this.#disposeAfter; } constructor(options) { - const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; + const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, onInsert, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; if (max !== 0 && !isPosInt(max)) { throw new TypeError("max option must be a nonnegative integer"); } @@ -99207,6 +99305,9 @@ var require_commonjs14 = __commonJS({ if (typeof dispose === "function") { this.#dispose = dispose; } + if (typeof onInsert === "function") { + this.#onInsert = onInsert; + } if (typeof disposeAfter === "function") { this.#disposeAfter = disposeAfter; this.#disposed = []; @@ -99215,6 +99316,7 @@ var require_commonjs14 = __commonJS({ this.#disposed = void 0; } this.#hasDispose = !!this.#dispose; + this.#hasOnInsert = !!this.#onInsert; this.#hasDisposeAfter = !!this.#disposeAfter; this.noDisposeOnSet = !!noDisposeOnSet; this.noUpdateTTL = !!noUpdateTTL; @@ -99617,7 +99719,7 @@ var require_commonjs14 = __commonJS({ } /** * Return an array of [key, {@link LRUCache.Entry}] tuples which can be - * passed to {@link LRLUCache#load}. + * passed to {@link LRUCache#load}. * * The `start` fields are calculated relative to a portable `Date.now()` * timestamp, even if `performance.now()` is available. @@ -99728,6 +99830,9 @@ var require_commonjs14 = __commonJS({ if (status) status.set = "add"; noUpdateTTL = false; + if (this.#hasOnInsert) { + this.#onInsert?.(v, k, "add"); + } } else { this.#moveToTail(index); const oldVal = this.#valList[index]; @@ -99763,6 +99868,9 @@ var require_commonjs14 = __commonJS({ } else if (status) { status.set = "update"; } + if (this.#hasOnInsert) { + this.onInsert?.(v, k, v === oldVal ? "update" : "replace"); + } } if (ttl !== 0 && !this.#ttls) { this.#initializeTTLTracking(); @@ -100288,7 +100396,7 @@ var require_commonjs14 = __commonJS({ }); // node_modules/minipass/dist/commonjs/index.js -var require_commonjs15 = __commonJS({ +var require_commonjs17 = __commonJS({ "node_modules/minipass/dist/commonjs/index.js"(exports2) { "use strict"; var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { @@ -101179,9 +101287,9 @@ var require_commonjs15 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js -var require_commonjs16 = __commonJS({ - "node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js"(exports2) { +// node_modules/path-scurry/dist/commonjs/index.js +var require_commonjs18 = __commonJS({ + "node_modules/path-scurry/dist/commonjs/index.js"(exports2) { "use strict"; var __createBinding4 = exports2 && exports2.__createBinding || (Object.create ? (function(o, m, k, k2) { if (k2 === void 0) k2 = k; @@ -101212,14 +101320,14 @@ var require_commonjs16 = __commonJS({ }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.PathScurry = exports2.Path = exports2.PathScurryDarwin = exports2.PathScurryPosix = exports2.PathScurryWin32 = exports2.PathScurryBase = exports2.PathPosix = exports2.PathWin32 = exports2.PathBase = exports2.ChildrenCache = exports2.ResolveCache = void 0; - var lru_cache_1 = require_commonjs14(); + var lru_cache_1 = require_commonjs16(); var node_path_1 = require("node:path"); var node_url_1 = require("node:url"); var fs_1 = require("fs"); var actualFS = __importStar4(require("node:fs")); var realpathSync = fs_1.realpathSync.native; var promises_1 = require("node:fs/promises"); - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var defaultFS = { lstatSync: fs_1.lstatSync, readdir: fs_1.readdir, @@ -101434,6 +101542,8 @@ var require_commonjs16 = __commonJS({ /** * Deprecated alias for Dirent['parentPath'] Somewhat counterintuitively, * this property refers to the *parent* path, not the path object itself. + * + * @deprecated */ get path() { return this.parentPath; @@ -102953,13 +103063,13 @@ var require_commonjs16 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js +// node_modules/glob/dist/commonjs/pattern.js var require_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js"(exports2) { + "node_modules/glob/dist/commonjs/pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Pattern = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var isPatternList = (pl) => pl.length >= 1; var isGlobList = (gl) => gl.length >= 1; var Pattern = class _Pattern { @@ -103127,13 +103237,13 @@ var require_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js +// node_modules/glob/dist/commonjs/ignore.js var require_ignore = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js"(exports2) { + "node_modules/glob/dist/commonjs/ignore.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Ignore = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var pattern_js_1 = require_pattern(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; var Ignore = class { @@ -103224,13 +103334,13 @@ var require_ignore = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js +// node_modules/glob/dist/commonjs/processor.js var require_processor = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js"(exports2) { + "node_modules/glob/dist/commonjs/processor.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Processor = exports2.SubWalks = exports2.MatchRecord = exports2.HasWalkedCache = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var HasWalkedCache = class _HasWalkedCache { store; constructor(store = /* @__PURE__ */ new Map()) { @@ -103457,13 +103567,13 @@ var require_processor = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js +// node_modules/glob/dist/commonjs/walker.js var require_walker = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js"(exports2) { + "node_modules/glob/dist/commonjs/walker.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.GlobStream = exports2.GlobWalker = exports2.GlobUtil = void 0; - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var ignore_js_1 = require_ignore(); var processor_js_1 = require_processor(); var makeIgnore = (ignore, opts) => typeof ignore === "string" ? new ignore_js_1.Ignore([ignore], opts) : Array.isArray(ignore) ? new ignore_js_1.Ignore(ignore, opts) : ignore; @@ -103797,15 +103907,15 @@ var require_walker = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js +// node_modules/glob/dist/commonjs/glob.js var require_glob2 = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js"(exports2) { + "node_modules/glob/dist/commonjs/glob.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Glob = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var node_url_1 = require("node:url"); - var path_scurry_1 = require_commonjs16(); + var path_scurry_1 = require_commonjs18(); var pattern_js_1 = require_pattern(); var walker_js_1 = require_walker(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; @@ -104010,13 +104120,13 @@ var require_glob2 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js +// node_modules/glob/dist/commonjs/has-magic.js var require_has_magic = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js"(exports2) { + "node_modules/glob/dist/commonjs/has-magic.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.hasMagic = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var hasMagic = (pattern, options = {}) => { if (!Array.isArray(pattern)) { pattern = [pattern]; @@ -104031,9 +104141,9 @@ var require_has_magic = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js -var require_commonjs17 = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js"(exports2) { +// node_modules/glob/dist/commonjs/index.js +var require_commonjs19 = __commonJS({ + "node_modules/glob/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.glob = exports2.sync = exports2.iterate = exports2.iterateSync = exports2.stream = exports2.streamSync = exports2.Ignore = exports2.hasMagic = exports2.Glob = exports2.unescape = exports2.escape = void 0; @@ -104042,10 +104152,10 @@ var require_commonjs17 = __commonJS({ exports2.globSync = globSync; exports2.globIterateSync = globIterateSync; exports2.globIterate = globIterate; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var glob_js_1 = require_glob2(); var has_magic_js_1 = require_has_magic(); - var minimatch_2 = require_commonjs13(); + var minimatch_2 = require_commonjs15(); Object.defineProperty(exports2, "escape", { enumerable: true, get: function() { return minimatch_2.escape; } }); @@ -104122,7 +104232,7 @@ var require_file3 = __commonJS({ var difference = require_difference(); var union = require_union(); var isPlainObject = require_isPlainObject(); - var glob2 = require_commonjs17(); + var glob2 = require_commonjs19(); var file = module2.exports = {}; var pathSeparatorRe = /[\/\\]/g; var processPatterns = function(patterns, fn) { diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 5dcfd5062..d37d43366 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -27694,14 +27694,14 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", - eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + eslint: "^8.57.1", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27725,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index 2655e4315..3049d15a2 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -27694,14 +27694,14 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", - eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + eslint: "^8.57.1", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27725,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } diff --git a/lib/init-action-post.js b/lib/init-action-post.js index c238f7d06..5dd89dcf6 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -27694,14 +27694,14 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", - eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + eslint: "^8.57.1", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27725,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -97357,52 +97358,128 @@ var require_isPlainObject = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/brace-expansion/index.js -var require_brace_expansion3 = __commonJS({ - "node_modules/archiver-utils/node_modules/brace-expansion/index.js"(exports2, module2) { - var balanced = require_balanced_match(); - module2.exports = expandTop; +// node_modules/@isaacs/balanced-match/dist/commonjs/index.js +var require_commonjs13 = __commonJS({ + "node_modules/@isaacs/balanced-match/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.range = exports2.balanced = void 0; + var balanced = (a, b, str2) => { + const ma = a instanceof RegExp ? maybeMatch(a, str2) : a; + const mb = b instanceof RegExp ? maybeMatch(b, str2) : b; + const r = ma !== null && mb != null && (0, exports2.range)(ma, mb, str2); + return r && { + start: r[0], + end: r[1], + pre: str2.slice(0, r[0]), + body: str2.slice(r[0] + ma.length, r[1]), + post: str2.slice(r[1] + mb.length) + }; + }; + exports2.balanced = balanced; + var maybeMatch = (reg, str2) => { + const m = str2.match(reg); + return m ? m[0] : null; + }; + var range = (a, b, str2) => { + let begs, beg, left, right = void 0, result; + let ai = str2.indexOf(a); + let bi = str2.indexOf(b, ai + 1); + let i = ai; + if (ai >= 0 && bi > 0) { + if (a === b) { + return [ai, bi]; + } + begs = []; + left = str2.length; + while (i >= 0 && !result) { + if (i === ai) { + begs.push(i); + ai = str2.indexOf(a, i + 1); + } else if (begs.length === 1) { + const r = begs.pop(); + if (r !== void 0) + result = [r, bi]; + } else { + beg = begs.pop(); + if (beg !== void 0 && beg < left) { + left = beg; + right = bi; + } + bi = str2.indexOf(b, i + 1); + } + i = ai < bi && ai >= 0 ? ai : bi; + } + if (begs.length && right !== void 0) { + result = [left, right]; + } + } + return result; + }; + exports2.range = range; + } +}); + +// node_modules/@isaacs/brace-expansion/dist/commonjs/index.js +var require_commonjs14 = __commonJS({ + "node_modules/@isaacs/brace-expansion/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.expand = expand; + var balanced_match_1 = require_commonjs13(); var escSlash = "\0SLASH" + Math.random() + "\0"; var escOpen = "\0OPEN" + Math.random() + "\0"; var escClose = "\0CLOSE" + Math.random() + "\0"; var escComma = "\0COMMA" + Math.random() + "\0"; var escPeriod = "\0PERIOD" + Math.random() + "\0"; + var escSlashPattern = new RegExp(escSlash, "g"); + var escOpenPattern = new RegExp(escOpen, "g"); + var escClosePattern = new RegExp(escClose, "g"); + var escCommaPattern = new RegExp(escComma, "g"); + var escPeriodPattern = new RegExp(escPeriod, "g"); + var slashPattern = /\\\\/g; + var openPattern = /\\{/g; + var closePattern = /\\}/g; + var commaPattern = /\\,/g; + var periodPattern = /\\./g; function numeric(str2) { - return parseInt(str2, 10) == str2 ? parseInt(str2, 10) : str2.charCodeAt(0); + return !isNaN(str2) ? parseInt(str2, 10) : str2.charCodeAt(0); } function escapeBraces(str2) { - return str2.split("\\\\").join(escSlash).split("\\{").join(escOpen).split("\\}").join(escClose).split("\\,").join(escComma).split("\\.").join(escPeriod); + return str2.replace(slashPattern, escSlash).replace(openPattern, escOpen).replace(closePattern, escClose).replace(commaPattern, escComma).replace(periodPattern, escPeriod); } function unescapeBraces(str2) { - return str2.split(escSlash).join("\\").split(escOpen).join("{").split(escClose).join("}").split(escComma).join(",").split(escPeriod).join("."); + return str2.replace(escSlashPattern, "\\").replace(escOpenPattern, "{").replace(escClosePattern, "}").replace(escCommaPattern, ",").replace(escPeriodPattern, "."); } function parseCommaParts(str2) { - if (!str2) + if (!str2) { return [""]; - var parts = []; - var m = balanced("{", "}", str2); - if (!m) + } + const parts = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) { return str2.split(","); - var pre = m.pre; - var body = m.body; - var post = m.post; - var p = pre.split(","); + } + const { pre, body, post } = m; + const p = pre.split(","); p[p.length - 1] += "{" + body + "}"; - var postParts = parseCommaParts(post); + const postParts = parseCommaParts(post); if (post.length) { + ; p[p.length - 1] += postParts.shift(); p.push.apply(p, postParts); } parts.push.apply(parts, p); return parts; } - function expandTop(str2) { - if (!str2) + function expand(str2) { + if (!str2) { return []; - if (str2.substr(0, 2) === "{}") { - str2 = "\\{\\}" + str2.substr(2); } - return expand(escapeBraces(str2), true).map(unescapeBraces); + if (str2.slice(0, 2) === "{}") { + str2 = "\\{\\}" + str2.slice(2); + } + return expand_(escapeBraces(str2), true).map(unescapeBraces); } function embrace(str2) { return "{" + str2 + "}"; @@ -97416,73 +97493,74 @@ var require_brace_expansion3 = __commonJS({ function gte5(i, y) { return i >= y; } - function expand(str2, isTop) { - var expansions = []; - var m = balanced("{", "}", str2); - if (!m) return [str2]; - var pre = m.pre; - var post = m.post.length ? expand(m.post, false) : [""]; + function expand_(str2, isTop) { + const expansions = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) + return [str2]; + const pre = m.pre; + const post = m.post.length ? expand_(m.post, false) : [""]; if (/\$$/.test(m.pre)) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + "{" + m.body + "}" + post[k]; + for (let k = 0; k < post.length; k++) { + const expansion = pre + "{" + m.body + "}" + post[k]; expansions.push(expansion); } } else { - var isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); - var isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); - var isSequence = isNumericSequence || isAlphaSequence; - var isOptions = m.body.indexOf(",") >= 0; + const isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); + const isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); + const isSequence = isNumericSequence || isAlphaSequence; + const isOptions = m.body.indexOf(",") >= 0; if (!isSequence && !isOptions) { if (m.post.match(/,(?!,).*\}/)) { str2 = m.pre + "{" + m.body + escClose + m.post; - return expand(str2); + return expand_(str2); } return [str2]; } - var n; + let n; if (isSequence) { n = m.body.split(/\.\./); } else { n = parseCommaParts(m.body); - if (n.length === 1) { - n = expand(n[0], false).map(embrace); + if (n.length === 1 && n[0] !== void 0) { + n = expand_(n[0], false).map(embrace); if (n.length === 1) { - return post.map(function(p) { - return m.pre + n[0] + p; - }); + return post.map((p) => m.pre + n[0] + p); } } } - var N; - if (isSequence) { - var x = numeric(n[0]); - var y = numeric(n[1]); - var width = Math.max(n[0].length, n[1].length); - var incr = n.length == 3 ? Math.abs(numeric(n[2])) : 1; - var test = lte; - var reverse = y < x; + let N; + if (isSequence && n[0] !== void 0 && n[1] !== void 0) { + const x = numeric(n[0]); + const y = numeric(n[1]); + const width = Math.max(n[0].length, n[1].length); + let incr = n.length === 3 && n[2] !== void 0 ? Math.abs(numeric(n[2])) : 1; + let test = lte; + const reverse = y < x; if (reverse) { incr *= -1; test = gte5; } - var pad = n.some(isPadded); + const pad = n.some(isPadded); N = []; - for (var i = x; test(i, y); i += incr) { - var c; + for (let i = x; test(i, y); i += incr) { + let c; if (isAlphaSequence) { c = String.fromCharCode(i); - if (c === "\\") + if (c === "\\") { c = ""; + } } else { c = String(i); if (pad) { - var need = width - c.length; + const need = width - c.length; if (need > 0) { - var z = new Array(need + 1).join("0"); - if (i < 0) + const z = new Array(need + 1).join("0"); + if (i < 0) { c = "-" + z + c.slice(1); - else + } else { c = z + c; + } } } } @@ -97490,15 +97568,16 @@ var require_brace_expansion3 = __commonJS({ } } else { N = []; - for (var j = 0; j < n.length; j++) { - N.push.apply(N, expand(n[j], false)); + for (let j = 0; j < n.length; j++) { + N.push.apply(N, expand_(n[j], false)); } } - for (var j = 0; j < N.length; j++) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + N[j] + post[k]; - if (!isTop || isSequence || expansion) + for (let j = 0; j < N.length; j++) { + for (let k = 0; k < post.length; k++) { + const expansion = pre + N[j] + post[k]; + if (!isTop || isSequence || expansion) { expansions.push(expansion); + } } } } @@ -97507,9 +97586,9 @@ var require_brace_expansion3 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js var require_assert_valid_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.assertValidPattern = void 0; @@ -97526,9 +97605,9 @@ var require_assert_valid_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js var require_brace_expressions = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.parseClass = void 0; @@ -97643,22 +97722,25 @@ var require_brace_expressions = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js var require_unescape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = void 0; - var unescape = (s, { windowsPathsNoEscape = false } = {}) => { - return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + var unescape = (s, { windowsPathsNoEscape = false, magicalBraces = true } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + } + return windowsPathsNoEscape ? s.replace(/\[([^\/\\{}])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\{}])\]/g, "$1$2").replace(/\\([^\/{}])/g, "$1"); }; exports2.unescape = unescape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js var require_ast = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.AST = void 0; @@ -98014,7 +98096,7 @@ var require_ast = __commonJS({ if (this.#root === this) this.#fillNegs(); if (!this.type) { - const noEmpty = this.isStart() && this.isEnd(); + const noEmpty = this.isStart() && this.isEnd() && !this.#parts.some((s) => typeof s !== "string"); const src = this.#parts.map((p) => { const [re, _2, hasMagic, uflag] = typeof p === "string" ? _AST.#parseGlob(p, this.#hasMagic, noEmpty) : p.toRegExpSource(allowDot); this.#hasMagic = this.#hasMagic || hasMagic; @@ -98124,10 +98206,7 @@ var require_ast = __commonJS({ } } if (c === "*") { - if (noEmpty && glob2 === "*") - re += starNoEmpty; - else - re += star; + re += noEmpty && glob2 === "*" ? starNoEmpty : star; hasMagic = true; continue; } @@ -98145,29 +98224,29 @@ var require_ast = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js var require_escape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.escape = void 0; - var escape = (s, { windowsPathsNoEscape = false } = {}) => { + var escape = (s, { windowsPathsNoEscape = false, magicalBraces = false } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/[?*()[\]{}]/g, "[$&]") : s.replace(/[?*()[\]\\{}]/g, "\\$&"); + } return windowsPathsNoEscape ? s.replace(/[?*()[\]]/g, "[$&]") : s.replace(/[?*()[\]\\]/g, "\\$&"); }; exports2.escape = escape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js -var require_commonjs13 = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js"(exports2) { +// node_modules/glob/node_modules/minimatch/dist/commonjs/index.js +var require_commonjs15 = __commonJS({ + "node_modules/glob/node_modules/minimatch/dist/commonjs/index.js"(exports2) { "use strict"; - var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { - return mod && mod.__esModule ? mod : { "default": mod }; - }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = exports2.escape = exports2.AST = exports2.Minimatch = exports2.match = exports2.makeRe = exports2.braceExpand = exports2.defaults = exports2.filter = exports2.GLOBSTAR = exports2.sep = exports2.minimatch = void 0; - var brace_expansion_1 = __importDefault4(require_brace_expansion3()); + var brace_expansion_1 = require_commonjs14(); var assert_valid_pattern_js_1 = require_assert_valid_pattern(); var ast_js_1 = require_ast(); var escape_js_1 = require_escape(); @@ -98290,7 +98369,7 @@ var require_commonjs13 = __commonJS({ if (options.nobrace || !/\{(?:(?!\{).)*\}/.test(pattern)) { return [pattern]; } - return (0, brace_expansion_1.default)(pattern); + return (0, brace_expansion_1.expand)(pattern); }; exports2.braceExpand = braceExpand; exports2.minimatch.braceExpand = exports2.braceExpand; @@ -98814,16 +98893,27 @@ var require_commonjs13 = __commonJS({ pp[i] = twoStar; } } else if (next === void 0) { - pp[i - 1] = prev + "(?:\\/|" + twoStar + ")?"; + pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + ")?"; } else if (next !== exports2.GLOBSTAR) { pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + "\\/)" + next; pp[i + 1] = exports2.GLOBSTAR; } }); - return pp.filter((p) => p !== exports2.GLOBSTAR).join("/"); + const filtered = pp.filter((p) => p !== exports2.GLOBSTAR); + if (this.partial && filtered.length >= 1) { + const prefixes = []; + for (let i = 1; i <= filtered.length; i++) { + prefixes.push(filtered.slice(0, i).join("/")); + } + return "(?:" + prefixes.join("|") + ")"; + } + return filtered.join("/"); }).join("|"); const [open, close] = set2.length > 1 ? ["(?:", ")"] : ["", ""]; re = "^" + open + re + close + "$"; + if (this.partial) { + re = "^(?:\\/|" + open + re.slice(1, -1) + close + ")$"; + } if (this.negate) re = "^(?!" + re + ").+$"; try { @@ -98910,9 +99000,9 @@ var require_commonjs13 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js -var require_commonjs14 = __commonJS({ - "node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js"(exports2) { +// node_modules/lru-cache/dist/commonjs/index.js +var require_commonjs16 = __commonJS({ + "node_modules/lru-cache/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.LRUCache = void 0; @@ -99001,6 +99091,7 @@ var require_commonjs14 = __commonJS({ #max; #maxSize; #dispose; + #onInsert; #disposeAfter; #fetchMethod; #memoMethod; @@ -99082,6 +99173,7 @@ var require_commonjs14 = __commonJS({ #hasDispose; #hasFetchMethod; #hasDisposeAfter; + #hasOnInsert; /** * Do not call this method unless you need to inspect the * inner workings of the cache. If anything returned by this @@ -99158,6 +99250,12 @@ var require_commonjs14 = __commonJS({ get dispose() { return this.#dispose; } + /** + * {@link LRUCache.OptionsBase.onInsert} (read-only) + */ + get onInsert() { + return this.#onInsert; + } /** * {@link LRUCache.OptionsBase.disposeAfter} (read-only) */ @@ -99165,7 +99263,7 @@ var require_commonjs14 = __commonJS({ return this.#disposeAfter; } constructor(options) { - const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; + const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, onInsert, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; if (max !== 0 && !isPosInt(max)) { throw new TypeError("max option must be a nonnegative integer"); } @@ -99207,6 +99305,9 @@ var require_commonjs14 = __commonJS({ if (typeof dispose === "function") { this.#dispose = dispose; } + if (typeof onInsert === "function") { + this.#onInsert = onInsert; + } if (typeof disposeAfter === "function") { this.#disposeAfter = disposeAfter; this.#disposed = []; @@ -99215,6 +99316,7 @@ var require_commonjs14 = __commonJS({ this.#disposed = void 0; } this.#hasDispose = !!this.#dispose; + this.#hasOnInsert = !!this.#onInsert; this.#hasDisposeAfter = !!this.#disposeAfter; this.noDisposeOnSet = !!noDisposeOnSet; this.noUpdateTTL = !!noUpdateTTL; @@ -99617,7 +99719,7 @@ var require_commonjs14 = __commonJS({ } /** * Return an array of [key, {@link LRUCache.Entry}] tuples which can be - * passed to {@link LRLUCache#load}. + * passed to {@link LRUCache#load}. * * The `start` fields are calculated relative to a portable `Date.now()` * timestamp, even if `performance.now()` is available. @@ -99728,6 +99830,9 @@ var require_commonjs14 = __commonJS({ if (status) status.set = "add"; noUpdateTTL = false; + if (this.#hasOnInsert) { + this.#onInsert?.(v, k, "add"); + } } else { this.#moveToTail(index); const oldVal = this.#valList[index]; @@ -99763,6 +99868,9 @@ var require_commonjs14 = __commonJS({ } else if (status) { status.set = "update"; } + if (this.#hasOnInsert) { + this.onInsert?.(v, k, v === oldVal ? "update" : "replace"); + } } if (ttl !== 0 && !this.#ttls) { this.#initializeTTLTracking(); @@ -100288,7 +100396,7 @@ var require_commonjs14 = __commonJS({ }); // node_modules/minipass/dist/commonjs/index.js -var require_commonjs15 = __commonJS({ +var require_commonjs17 = __commonJS({ "node_modules/minipass/dist/commonjs/index.js"(exports2) { "use strict"; var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { @@ -101179,9 +101287,9 @@ var require_commonjs15 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js -var require_commonjs16 = __commonJS({ - "node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js"(exports2) { +// node_modules/path-scurry/dist/commonjs/index.js +var require_commonjs18 = __commonJS({ + "node_modules/path-scurry/dist/commonjs/index.js"(exports2) { "use strict"; var __createBinding4 = exports2 && exports2.__createBinding || (Object.create ? (function(o, m, k, k2) { if (k2 === void 0) k2 = k; @@ -101212,14 +101320,14 @@ var require_commonjs16 = __commonJS({ }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.PathScurry = exports2.Path = exports2.PathScurryDarwin = exports2.PathScurryPosix = exports2.PathScurryWin32 = exports2.PathScurryBase = exports2.PathPosix = exports2.PathWin32 = exports2.PathBase = exports2.ChildrenCache = exports2.ResolveCache = void 0; - var lru_cache_1 = require_commonjs14(); + var lru_cache_1 = require_commonjs16(); var node_path_1 = require("node:path"); var node_url_1 = require("node:url"); var fs_1 = require("fs"); var actualFS = __importStar4(require("node:fs")); var realpathSync = fs_1.realpathSync.native; var promises_1 = require("node:fs/promises"); - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var defaultFS = { lstatSync: fs_1.lstatSync, readdir: fs_1.readdir, @@ -101434,6 +101542,8 @@ var require_commonjs16 = __commonJS({ /** * Deprecated alias for Dirent['parentPath'] Somewhat counterintuitively, * this property refers to the *parent* path, not the path object itself. + * + * @deprecated */ get path() { return this.parentPath; @@ -102953,13 +103063,13 @@ var require_commonjs16 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js +// node_modules/glob/dist/commonjs/pattern.js var require_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js"(exports2) { + "node_modules/glob/dist/commonjs/pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Pattern = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var isPatternList = (pl) => pl.length >= 1; var isGlobList = (gl) => gl.length >= 1; var Pattern = class _Pattern { @@ -103127,13 +103237,13 @@ var require_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js +// node_modules/glob/dist/commonjs/ignore.js var require_ignore = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js"(exports2) { + "node_modules/glob/dist/commonjs/ignore.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Ignore = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var pattern_js_1 = require_pattern(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; var Ignore = class { @@ -103224,13 +103334,13 @@ var require_ignore = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js +// node_modules/glob/dist/commonjs/processor.js var require_processor = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js"(exports2) { + "node_modules/glob/dist/commonjs/processor.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Processor = exports2.SubWalks = exports2.MatchRecord = exports2.HasWalkedCache = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var HasWalkedCache = class _HasWalkedCache { store; constructor(store = /* @__PURE__ */ new Map()) { @@ -103457,13 +103567,13 @@ var require_processor = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js +// node_modules/glob/dist/commonjs/walker.js var require_walker = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js"(exports2) { + "node_modules/glob/dist/commonjs/walker.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.GlobStream = exports2.GlobWalker = exports2.GlobUtil = void 0; - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var ignore_js_1 = require_ignore(); var processor_js_1 = require_processor(); var makeIgnore = (ignore, opts) => typeof ignore === "string" ? new ignore_js_1.Ignore([ignore], opts) : Array.isArray(ignore) ? new ignore_js_1.Ignore(ignore, opts) : ignore; @@ -103797,15 +103907,15 @@ var require_walker = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js +// node_modules/glob/dist/commonjs/glob.js var require_glob2 = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js"(exports2) { + "node_modules/glob/dist/commonjs/glob.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Glob = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var node_url_1 = require("node:url"); - var path_scurry_1 = require_commonjs16(); + var path_scurry_1 = require_commonjs18(); var pattern_js_1 = require_pattern(); var walker_js_1 = require_walker(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; @@ -104010,13 +104120,13 @@ var require_glob2 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js +// node_modules/glob/dist/commonjs/has-magic.js var require_has_magic = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js"(exports2) { + "node_modules/glob/dist/commonjs/has-magic.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.hasMagic = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var hasMagic = (pattern, options = {}) => { if (!Array.isArray(pattern)) { pattern = [pattern]; @@ -104031,9 +104141,9 @@ var require_has_magic = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js -var require_commonjs17 = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js"(exports2) { +// node_modules/glob/dist/commonjs/index.js +var require_commonjs19 = __commonJS({ + "node_modules/glob/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.glob = exports2.sync = exports2.iterate = exports2.iterateSync = exports2.stream = exports2.streamSync = exports2.Ignore = exports2.hasMagic = exports2.Glob = exports2.unescape = exports2.escape = void 0; @@ -104042,10 +104152,10 @@ var require_commonjs17 = __commonJS({ exports2.globSync = globSync; exports2.globIterateSync = globIterateSync; exports2.globIterate = globIterate; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var glob_js_1 = require_glob2(); var has_magic_js_1 = require_has_magic(); - var minimatch_2 = require_commonjs13(); + var minimatch_2 = require_commonjs15(); Object.defineProperty(exports2, "escape", { enumerable: true, get: function() { return minimatch_2.escape; } }); @@ -104122,7 +104232,7 @@ var require_file3 = __commonJS({ var difference = require_difference(); var union = require_union(); var isPlainObject = require_isPlainObject(); - var glob2 = require_commonjs17(); + var glob2 = require_commonjs19(); var file = module2.exports = {}; var pathSeparatorRe = /[\/\\]/g; var processPatterns = function(patterns, fn) { diff --git a/lib/init-action.js b/lib/init-action.js index 6de0e4e2d..c3f536a33 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -27694,14 +27694,14 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", - eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + eslint: "^8.57.1", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27725,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index 8a757d8c6..c3d54f680 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -27694,14 +27694,14 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", - eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + eslint: "^8.57.1", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27725,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index 31f135f53..973e9c431 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -27694,14 +27694,14 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", - eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + eslint: "^8.57.1", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27725,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index 86991078a..7e34a5e95 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -27694,14 +27694,14 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", - eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + eslint: "^8.57.1", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27725,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -94132,52 +94133,128 @@ var require_isPlainObject = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/brace-expansion/index.js -var require_brace_expansion3 = __commonJS({ - "node_modules/archiver-utils/node_modules/brace-expansion/index.js"(exports2, module2) { - var balanced = require_balanced_match(); - module2.exports = expandTop; +// node_modules/@isaacs/balanced-match/dist/commonjs/index.js +var require_commonjs13 = __commonJS({ + "node_modules/@isaacs/balanced-match/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.range = exports2.balanced = void 0; + var balanced = (a, b, str2) => { + const ma = a instanceof RegExp ? maybeMatch(a, str2) : a; + const mb = b instanceof RegExp ? maybeMatch(b, str2) : b; + const r = ma !== null && mb != null && (0, exports2.range)(ma, mb, str2); + return r && { + start: r[0], + end: r[1], + pre: str2.slice(0, r[0]), + body: str2.slice(r[0] + ma.length, r[1]), + post: str2.slice(r[1] + mb.length) + }; + }; + exports2.balanced = balanced; + var maybeMatch = (reg, str2) => { + const m = str2.match(reg); + return m ? m[0] : null; + }; + var range = (a, b, str2) => { + let begs, beg, left, right = void 0, result; + let ai = str2.indexOf(a); + let bi = str2.indexOf(b, ai + 1); + let i = ai; + if (ai >= 0 && bi > 0) { + if (a === b) { + return [ai, bi]; + } + begs = []; + left = str2.length; + while (i >= 0 && !result) { + if (i === ai) { + begs.push(i); + ai = str2.indexOf(a, i + 1); + } else if (begs.length === 1) { + const r = begs.pop(); + if (r !== void 0) + result = [r, bi]; + } else { + beg = begs.pop(); + if (beg !== void 0 && beg < left) { + left = beg; + right = bi; + } + bi = str2.indexOf(b, i + 1); + } + i = ai < bi && ai >= 0 ? ai : bi; + } + if (begs.length && right !== void 0) { + result = [left, right]; + } + } + return result; + }; + exports2.range = range; + } +}); + +// node_modules/@isaacs/brace-expansion/dist/commonjs/index.js +var require_commonjs14 = __commonJS({ + "node_modules/@isaacs/brace-expansion/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.expand = expand; + var balanced_match_1 = require_commonjs13(); var escSlash = "\0SLASH" + Math.random() + "\0"; var escOpen = "\0OPEN" + Math.random() + "\0"; var escClose = "\0CLOSE" + Math.random() + "\0"; var escComma = "\0COMMA" + Math.random() + "\0"; var escPeriod = "\0PERIOD" + Math.random() + "\0"; + var escSlashPattern = new RegExp(escSlash, "g"); + var escOpenPattern = new RegExp(escOpen, "g"); + var escClosePattern = new RegExp(escClose, "g"); + var escCommaPattern = new RegExp(escComma, "g"); + var escPeriodPattern = new RegExp(escPeriod, "g"); + var slashPattern = /\\\\/g; + var openPattern = /\\{/g; + var closePattern = /\\}/g; + var commaPattern = /\\,/g; + var periodPattern = /\\./g; function numeric(str2) { - return parseInt(str2, 10) == str2 ? parseInt(str2, 10) : str2.charCodeAt(0); + return !isNaN(str2) ? parseInt(str2, 10) : str2.charCodeAt(0); } function escapeBraces(str2) { - return str2.split("\\\\").join(escSlash).split("\\{").join(escOpen).split("\\}").join(escClose).split("\\,").join(escComma).split("\\.").join(escPeriod); + return str2.replace(slashPattern, escSlash).replace(openPattern, escOpen).replace(closePattern, escClose).replace(commaPattern, escComma).replace(periodPattern, escPeriod); } function unescapeBraces(str2) { - return str2.split(escSlash).join("\\").split(escOpen).join("{").split(escClose).join("}").split(escComma).join(",").split(escPeriod).join("."); + return str2.replace(escSlashPattern, "\\").replace(escOpenPattern, "{").replace(escClosePattern, "}").replace(escCommaPattern, ",").replace(escPeriodPattern, "."); } function parseCommaParts(str2) { - if (!str2) + if (!str2) { return [""]; - var parts = []; - var m = balanced("{", "}", str2); - if (!m) + } + const parts = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) { return str2.split(","); - var pre = m.pre; - var body = m.body; - var post = m.post; - var p = pre.split(","); + } + const { pre, body, post } = m; + const p = pre.split(","); p[p.length - 1] += "{" + body + "}"; - var postParts = parseCommaParts(post); + const postParts = parseCommaParts(post); if (post.length) { + ; p[p.length - 1] += postParts.shift(); p.push.apply(p, postParts); } parts.push.apply(parts, p); return parts; } - function expandTop(str2) { - if (!str2) + function expand(str2) { + if (!str2) { return []; - if (str2.substr(0, 2) === "{}") { - str2 = "\\{\\}" + str2.substr(2); } - return expand(escapeBraces(str2), true).map(unescapeBraces); + if (str2.slice(0, 2) === "{}") { + str2 = "\\{\\}" + str2.slice(2); + } + return expand_(escapeBraces(str2), true).map(unescapeBraces); } function embrace(str2) { return "{" + str2 + "}"; @@ -94191,73 +94268,74 @@ var require_brace_expansion3 = __commonJS({ function gte5(i, y) { return i >= y; } - function expand(str2, isTop) { - var expansions = []; - var m = balanced("{", "}", str2); - if (!m) return [str2]; - var pre = m.pre; - var post = m.post.length ? expand(m.post, false) : [""]; + function expand_(str2, isTop) { + const expansions = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) + return [str2]; + const pre = m.pre; + const post = m.post.length ? expand_(m.post, false) : [""]; if (/\$$/.test(m.pre)) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + "{" + m.body + "}" + post[k]; + for (let k = 0; k < post.length; k++) { + const expansion = pre + "{" + m.body + "}" + post[k]; expansions.push(expansion); } } else { - var isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); - var isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); - var isSequence = isNumericSequence || isAlphaSequence; - var isOptions = m.body.indexOf(",") >= 0; + const isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); + const isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); + const isSequence = isNumericSequence || isAlphaSequence; + const isOptions = m.body.indexOf(",") >= 0; if (!isSequence && !isOptions) { if (m.post.match(/,(?!,).*\}/)) { str2 = m.pre + "{" + m.body + escClose + m.post; - return expand(str2); + return expand_(str2); } return [str2]; } - var n; + let n; if (isSequence) { n = m.body.split(/\.\./); } else { n = parseCommaParts(m.body); - if (n.length === 1) { - n = expand(n[0], false).map(embrace); + if (n.length === 1 && n[0] !== void 0) { + n = expand_(n[0], false).map(embrace); if (n.length === 1) { - return post.map(function(p) { - return m.pre + n[0] + p; - }); + return post.map((p) => m.pre + n[0] + p); } } } - var N; - if (isSequence) { - var x = numeric(n[0]); - var y = numeric(n[1]); - var width = Math.max(n[0].length, n[1].length); - var incr = n.length == 3 ? Math.abs(numeric(n[2])) : 1; - var test = lte; - var reverse = y < x; + let N; + if (isSequence && n[0] !== void 0 && n[1] !== void 0) { + const x = numeric(n[0]); + const y = numeric(n[1]); + const width = Math.max(n[0].length, n[1].length); + let incr = n.length === 3 && n[2] !== void 0 ? Math.abs(numeric(n[2])) : 1; + let test = lte; + const reverse = y < x; if (reverse) { incr *= -1; test = gte5; } - var pad = n.some(isPadded); + const pad = n.some(isPadded); N = []; - for (var i = x; test(i, y); i += incr) { - var c; + for (let i = x; test(i, y); i += incr) { + let c; if (isAlphaSequence) { c = String.fromCharCode(i); - if (c === "\\") + if (c === "\\") { c = ""; + } } else { c = String(i); if (pad) { - var need = width - c.length; + const need = width - c.length; if (need > 0) { - var z = new Array(need + 1).join("0"); - if (i < 0) + const z = new Array(need + 1).join("0"); + if (i < 0) { c = "-" + z + c.slice(1); - else + } else { c = z + c; + } } } } @@ -94265,15 +94343,16 @@ var require_brace_expansion3 = __commonJS({ } } else { N = []; - for (var j = 0; j < n.length; j++) { - N.push.apply(N, expand(n[j], false)); + for (let j = 0; j < n.length; j++) { + N.push.apply(N, expand_(n[j], false)); } } - for (var j = 0; j < N.length; j++) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + N[j] + post[k]; - if (!isTop || isSequence || expansion) + for (let j = 0; j < N.length; j++) { + for (let k = 0; k < post.length; k++) { + const expansion = pre + N[j] + post[k]; + if (!isTop || isSequence || expansion) { expansions.push(expansion); + } } } } @@ -94282,9 +94361,9 @@ var require_brace_expansion3 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js var require_assert_valid_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.assertValidPattern = void 0; @@ -94301,9 +94380,9 @@ var require_assert_valid_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js var require_brace_expressions = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.parseClass = void 0; @@ -94418,22 +94497,25 @@ var require_brace_expressions = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js var require_unescape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = void 0; - var unescape = (s, { windowsPathsNoEscape = false } = {}) => { - return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + var unescape = (s, { windowsPathsNoEscape = false, magicalBraces = true } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + } + return windowsPathsNoEscape ? s.replace(/\[([^\/\\{}])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\{}])\]/g, "$1$2").replace(/\\([^\/{}])/g, "$1"); }; exports2.unescape = unescape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js var require_ast = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.AST = void 0; @@ -94789,7 +94871,7 @@ var require_ast = __commonJS({ if (this.#root === this) this.#fillNegs(); if (!this.type) { - const noEmpty = this.isStart() && this.isEnd(); + const noEmpty = this.isStart() && this.isEnd() && !this.#parts.some((s) => typeof s !== "string"); const src = this.#parts.map((p) => { const [re, _2, hasMagic, uflag] = typeof p === "string" ? _AST.#parseGlob(p, this.#hasMagic, noEmpty) : p.toRegExpSource(allowDot); this.#hasMagic = this.#hasMagic || hasMagic; @@ -94899,10 +94981,7 @@ var require_ast = __commonJS({ } } if (c === "*") { - if (noEmpty && glob2 === "*") - re += starNoEmpty; - else - re += star; + re += noEmpty && glob2 === "*" ? starNoEmpty : star; hasMagic = true; continue; } @@ -94920,29 +94999,29 @@ var require_ast = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js var require_escape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.escape = void 0; - var escape = (s, { windowsPathsNoEscape = false } = {}) => { + var escape = (s, { windowsPathsNoEscape = false, magicalBraces = false } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/[?*()[\]{}]/g, "[$&]") : s.replace(/[?*()[\]\\{}]/g, "\\$&"); + } return windowsPathsNoEscape ? s.replace(/[?*()[\]]/g, "[$&]") : s.replace(/[?*()[\]\\]/g, "\\$&"); }; exports2.escape = escape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js -var require_commonjs13 = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js"(exports2) { +// node_modules/glob/node_modules/minimatch/dist/commonjs/index.js +var require_commonjs15 = __commonJS({ + "node_modules/glob/node_modules/minimatch/dist/commonjs/index.js"(exports2) { "use strict"; - var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { - return mod && mod.__esModule ? mod : { "default": mod }; - }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = exports2.escape = exports2.AST = exports2.Minimatch = exports2.match = exports2.makeRe = exports2.braceExpand = exports2.defaults = exports2.filter = exports2.GLOBSTAR = exports2.sep = exports2.minimatch = void 0; - var brace_expansion_1 = __importDefault4(require_brace_expansion3()); + var brace_expansion_1 = require_commonjs14(); var assert_valid_pattern_js_1 = require_assert_valid_pattern(); var ast_js_1 = require_ast(); var escape_js_1 = require_escape(); @@ -95065,7 +95144,7 @@ var require_commonjs13 = __commonJS({ if (options.nobrace || !/\{(?:(?!\{).)*\}/.test(pattern)) { return [pattern]; } - return (0, brace_expansion_1.default)(pattern); + return (0, brace_expansion_1.expand)(pattern); }; exports2.braceExpand = braceExpand; exports2.minimatch.braceExpand = exports2.braceExpand; @@ -95589,16 +95668,27 @@ var require_commonjs13 = __commonJS({ pp[i] = twoStar; } } else if (next === void 0) { - pp[i - 1] = prev + "(?:\\/|" + twoStar + ")?"; + pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + ")?"; } else if (next !== exports2.GLOBSTAR) { pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + "\\/)" + next; pp[i + 1] = exports2.GLOBSTAR; } }); - return pp.filter((p) => p !== exports2.GLOBSTAR).join("/"); + const filtered = pp.filter((p) => p !== exports2.GLOBSTAR); + if (this.partial && filtered.length >= 1) { + const prefixes = []; + for (let i = 1; i <= filtered.length; i++) { + prefixes.push(filtered.slice(0, i).join("/")); + } + return "(?:" + prefixes.join("|") + ")"; + } + return filtered.join("/"); }).join("|"); const [open, close] = set2.length > 1 ? ["(?:", ")"] : ["", ""]; re = "^" + open + re + close + "$"; + if (this.partial) { + re = "^(?:\\/|" + open + re.slice(1, -1) + close + ")$"; + } if (this.negate) re = "^(?!" + re + ").+$"; try { @@ -95685,9 +95775,9 @@ var require_commonjs13 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js -var require_commonjs14 = __commonJS({ - "node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js"(exports2) { +// node_modules/lru-cache/dist/commonjs/index.js +var require_commonjs16 = __commonJS({ + "node_modules/lru-cache/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.LRUCache = void 0; @@ -95776,6 +95866,7 @@ var require_commonjs14 = __commonJS({ #max; #maxSize; #dispose; + #onInsert; #disposeAfter; #fetchMethod; #memoMethod; @@ -95857,6 +95948,7 @@ var require_commonjs14 = __commonJS({ #hasDispose; #hasFetchMethod; #hasDisposeAfter; + #hasOnInsert; /** * Do not call this method unless you need to inspect the * inner workings of the cache. If anything returned by this @@ -95933,6 +96025,12 @@ var require_commonjs14 = __commonJS({ get dispose() { return this.#dispose; } + /** + * {@link LRUCache.OptionsBase.onInsert} (read-only) + */ + get onInsert() { + return this.#onInsert; + } /** * {@link LRUCache.OptionsBase.disposeAfter} (read-only) */ @@ -95940,7 +96038,7 @@ var require_commonjs14 = __commonJS({ return this.#disposeAfter; } constructor(options) { - const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; + const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, onInsert, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; if (max !== 0 && !isPosInt(max)) { throw new TypeError("max option must be a nonnegative integer"); } @@ -95982,6 +96080,9 @@ var require_commonjs14 = __commonJS({ if (typeof dispose === "function") { this.#dispose = dispose; } + if (typeof onInsert === "function") { + this.#onInsert = onInsert; + } if (typeof disposeAfter === "function") { this.#disposeAfter = disposeAfter; this.#disposed = []; @@ -95990,6 +96091,7 @@ var require_commonjs14 = __commonJS({ this.#disposed = void 0; } this.#hasDispose = !!this.#dispose; + this.#hasOnInsert = !!this.#onInsert; this.#hasDisposeAfter = !!this.#disposeAfter; this.noDisposeOnSet = !!noDisposeOnSet; this.noUpdateTTL = !!noUpdateTTL; @@ -96392,7 +96494,7 @@ var require_commonjs14 = __commonJS({ } /** * Return an array of [key, {@link LRUCache.Entry}] tuples which can be - * passed to {@link LRLUCache#load}. + * passed to {@link LRUCache#load}. * * The `start` fields are calculated relative to a portable `Date.now()` * timestamp, even if `performance.now()` is available. @@ -96503,6 +96605,9 @@ var require_commonjs14 = __commonJS({ if (status) status.set = "add"; noUpdateTTL = false; + if (this.#hasOnInsert) { + this.#onInsert?.(v, k, "add"); + } } else { this.#moveToTail(index); const oldVal = this.#valList[index]; @@ -96538,6 +96643,9 @@ var require_commonjs14 = __commonJS({ } else if (status) { status.set = "update"; } + if (this.#hasOnInsert) { + this.onInsert?.(v, k, v === oldVal ? "update" : "replace"); + } } if (ttl !== 0 && !this.#ttls) { this.#initializeTTLTracking(); @@ -97063,7 +97171,7 @@ var require_commonjs14 = __commonJS({ }); // node_modules/minipass/dist/commonjs/index.js -var require_commonjs15 = __commonJS({ +var require_commonjs17 = __commonJS({ "node_modules/minipass/dist/commonjs/index.js"(exports2) { "use strict"; var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { @@ -97954,9 +98062,9 @@ var require_commonjs15 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js -var require_commonjs16 = __commonJS({ - "node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js"(exports2) { +// node_modules/path-scurry/dist/commonjs/index.js +var require_commonjs18 = __commonJS({ + "node_modules/path-scurry/dist/commonjs/index.js"(exports2) { "use strict"; var __createBinding4 = exports2 && exports2.__createBinding || (Object.create ? (function(o, m, k, k2) { if (k2 === void 0) k2 = k; @@ -97987,14 +98095,14 @@ var require_commonjs16 = __commonJS({ }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.PathScurry = exports2.Path = exports2.PathScurryDarwin = exports2.PathScurryPosix = exports2.PathScurryWin32 = exports2.PathScurryBase = exports2.PathPosix = exports2.PathWin32 = exports2.PathBase = exports2.ChildrenCache = exports2.ResolveCache = void 0; - var lru_cache_1 = require_commonjs14(); + var lru_cache_1 = require_commonjs16(); var node_path_1 = require("node:path"); var node_url_1 = require("node:url"); var fs_1 = require("fs"); var actualFS = __importStar4(require("node:fs")); var realpathSync = fs_1.realpathSync.native; var promises_1 = require("node:fs/promises"); - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var defaultFS = { lstatSync: fs_1.lstatSync, readdir: fs_1.readdir, @@ -98209,6 +98317,8 @@ var require_commonjs16 = __commonJS({ /** * Deprecated alias for Dirent['parentPath'] Somewhat counterintuitively, * this property refers to the *parent* path, not the path object itself. + * + * @deprecated */ get path() { return this.parentPath; @@ -99728,13 +99838,13 @@ var require_commonjs16 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js +// node_modules/glob/dist/commonjs/pattern.js var require_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js"(exports2) { + "node_modules/glob/dist/commonjs/pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Pattern = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var isPatternList = (pl) => pl.length >= 1; var isGlobList = (gl) => gl.length >= 1; var Pattern = class _Pattern { @@ -99902,13 +100012,13 @@ var require_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js +// node_modules/glob/dist/commonjs/ignore.js var require_ignore = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js"(exports2) { + "node_modules/glob/dist/commonjs/ignore.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Ignore = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var pattern_js_1 = require_pattern(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; var Ignore = class { @@ -99999,13 +100109,13 @@ var require_ignore = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js +// node_modules/glob/dist/commonjs/processor.js var require_processor = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js"(exports2) { + "node_modules/glob/dist/commonjs/processor.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Processor = exports2.SubWalks = exports2.MatchRecord = exports2.HasWalkedCache = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var HasWalkedCache = class _HasWalkedCache { store; constructor(store = /* @__PURE__ */ new Map()) { @@ -100232,13 +100342,13 @@ var require_processor = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js +// node_modules/glob/dist/commonjs/walker.js var require_walker = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js"(exports2) { + "node_modules/glob/dist/commonjs/walker.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.GlobStream = exports2.GlobWalker = exports2.GlobUtil = void 0; - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var ignore_js_1 = require_ignore(); var processor_js_1 = require_processor(); var makeIgnore = (ignore, opts) => typeof ignore === "string" ? new ignore_js_1.Ignore([ignore], opts) : Array.isArray(ignore) ? new ignore_js_1.Ignore(ignore, opts) : ignore; @@ -100572,15 +100682,15 @@ var require_walker = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js +// node_modules/glob/dist/commonjs/glob.js var require_glob2 = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js"(exports2) { + "node_modules/glob/dist/commonjs/glob.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Glob = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var node_url_1 = require("node:url"); - var path_scurry_1 = require_commonjs16(); + var path_scurry_1 = require_commonjs18(); var pattern_js_1 = require_pattern(); var walker_js_1 = require_walker(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; @@ -100785,13 +100895,13 @@ var require_glob2 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js +// node_modules/glob/dist/commonjs/has-magic.js var require_has_magic = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js"(exports2) { + "node_modules/glob/dist/commonjs/has-magic.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.hasMagic = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var hasMagic = (pattern, options = {}) => { if (!Array.isArray(pattern)) { pattern = [pattern]; @@ -100806,9 +100916,9 @@ var require_has_magic = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js -var require_commonjs17 = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js"(exports2) { +// node_modules/glob/dist/commonjs/index.js +var require_commonjs19 = __commonJS({ + "node_modules/glob/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.glob = exports2.sync = exports2.iterate = exports2.iterateSync = exports2.stream = exports2.streamSync = exports2.Ignore = exports2.hasMagic = exports2.Glob = exports2.unescape = exports2.escape = void 0; @@ -100817,10 +100927,10 @@ var require_commonjs17 = __commonJS({ exports2.globSync = globSync; exports2.globIterateSync = globIterateSync; exports2.globIterate = globIterate; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var glob_js_1 = require_glob2(); var has_magic_js_1 = require_has_magic(); - var minimatch_2 = require_commonjs13(); + var minimatch_2 = require_commonjs15(); Object.defineProperty(exports2, "escape", { enumerable: true, get: function() { return minimatch_2.escape; } }); @@ -100897,7 +101007,7 @@ var require_file3 = __commonJS({ var difference = require_difference(); var union = require_union(); var isPlainObject = require_isPlainObject(); - var glob2 = require_commonjs17(); + var glob2 = require_commonjs19(); var file = module2.exports = {}; var pathSeparatorRe = /[\/\\]/g; var processPatterns = function(patterns, fn) { diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index 3b6d4deca..c0869dd96 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -47352,14 +47352,14 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", - eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + eslint: "^8.57.1", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -47383,7 +47383,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } diff --git a/lib/upload-lib.js b/lib/upload-lib.js index e9e0a6747..ea6e2ca41 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -28991,14 +28991,14 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", - eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + eslint: "^8.57.1", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -29022,7 +29022,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index 96a6bc64b..fce0c4f79 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -27694,14 +27694,14 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", - eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + eslint: "^8.57.1", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27725,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -85693,52 +85694,128 @@ var require_isPlainObject = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/brace-expansion/index.js -var require_brace_expansion2 = __commonJS({ - "node_modules/archiver-utils/node_modules/brace-expansion/index.js"(exports2, module2) { - var balanced = require_balanced_match(); - module2.exports = expandTop; +// node_modules/@isaacs/balanced-match/dist/commonjs/index.js +var require_commonjs13 = __commonJS({ + "node_modules/@isaacs/balanced-match/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.range = exports2.balanced = void 0; + var balanced = (a, b, str2) => { + const ma = a instanceof RegExp ? maybeMatch(a, str2) : a; + const mb = b instanceof RegExp ? maybeMatch(b, str2) : b; + const r = ma !== null && mb != null && (0, exports2.range)(ma, mb, str2); + return r && { + start: r[0], + end: r[1], + pre: str2.slice(0, r[0]), + body: str2.slice(r[0] + ma.length, r[1]), + post: str2.slice(r[1] + mb.length) + }; + }; + exports2.balanced = balanced; + var maybeMatch = (reg, str2) => { + const m = str2.match(reg); + return m ? m[0] : null; + }; + var range = (a, b, str2) => { + let begs, beg, left, right = void 0, result; + let ai = str2.indexOf(a); + let bi = str2.indexOf(b, ai + 1); + let i = ai; + if (ai >= 0 && bi > 0) { + if (a === b) { + return [ai, bi]; + } + begs = []; + left = str2.length; + while (i >= 0 && !result) { + if (i === ai) { + begs.push(i); + ai = str2.indexOf(a, i + 1); + } else if (begs.length === 1) { + const r = begs.pop(); + if (r !== void 0) + result = [r, bi]; + } else { + beg = begs.pop(); + if (beg !== void 0 && beg < left) { + left = beg; + right = bi; + } + bi = str2.indexOf(b, i + 1); + } + i = ai < bi && ai >= 0 ? ai : bi; + } + if (begs.length && right !== void 0) { + result = [left, right]; + } + } + return result; + }; + exports2.range = range; + } +}); + +// node_modules/@isaacs/brace-expansion/dist/commonjs/index.js +var require_commonjs14 = __commonJS({ + "node_modules/@isaacs/brace-expansion/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.expand = expand; + var balanced_match_1 = require_commonjs13(); var escSlash = "\0SLASH" + Math.random() + "\0"; var escOpen = "\0OPEN" + Math.random() + "\0"; var escClose = "\0CLOSE" + Math.random() + "\0"; var escComma = "\0COMMA" + Math.random() + "\0"; var escPeriod = "\0PERIOD" + Math.random() + "\0"; + var escSlashPattern = new RegExp(escSlash, "g"); + var escOpenPattern = new RegExp(escOpen, "g"); + var escClosePattern = new RegExp(escClose, "g"); + var escCommaPattern = new RegExp(escComma, "g"); + var escPeriodPattern = new RegExp(escPeriod, "g"); + var slashPattern = /\\\\/g; + var openPattern = /\\{/g; + var closePattern = /\\}/g; + var commaPattern = /\\,/g; + var periodPattern = /\\./g; function numeric(str2) { - return parseInt(str2, 10) == str2 ? parseInt(str2, 10) : str2.charCodeAt(0); + return !isNaN(str2) ? parseInt(str2, 10) : str2.charCodeAt(0); } function escapeBraces(str2) { - return str2.split("\\\\").join(escSlash).split("\\{").join(escOpen).split("\\}").join(escClose).split("\\,").join(escComma).split("\\.").join(escPeriod); + return str2.replace(slashPattern, escSlash).replace(openPattern, escOpen).replace(closePattern, escClose).replace(commaPattern, escComma).replace(periodPattern, escPeriod); } function unescapeBraces(str2) { - return str2.split(escSlash).join("\\").split(escOpen).join("{").split(escClose).join("}").split(escComma).join(",").split(escPeriod).join("."); + return str2.replace(escSlashPattern, "\\").replace(escOpenPattern, "{").replace(escClosePattern, "}").replace(escCommaPattern, ",").replace(escPeriodPattern, "."); } function parseCommaParts(str2) { - if (!str2) + if (!str2) { return [""]; - var parts = []; - var m = balanced("{", "}", str2); - if (!m) + } + const parts = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) { return str2.split(","); - var pre = m.pre; - var body = m.body; - var post = m.post; - var p = pre.split(","); + } + const { pre, body, post } = m; + const p = pre.split(","); p[p.length - 1] += "{" + body + "}"; - var postParts = parseCommaParts(post); + const postParts = parseCommaParts(post); if (post.length) { + ; p[p.length - 1] += postParts.shift(); p.push.apply(p, postParts); } parts.push.apply(parts, p); return parts; } - function expandTop(str2) { - if (!str2) + function expand(str2) { + if (!str2) { return []; - if (str2.substr(0, 2) === "{}") { - str2 = "\\{\\}" + str2.substr(2); } - return expand(escapeBraces(str2), true).map(unescapeBraces); + if (str2.slice(0, 2) === "{}") { + str2 = "\\{\\}" + str2.slice(2); + } + return expand_(escapeBraces(str2), true).map(unescapeBraces); } function embrace(str2) { return "{" + str2 + "}"; @@ -85752,73 +85829,74 @@ var require_brace_expansion2 = __commonJS({ function gte5(i, y) { return i >= y; } - function expand(str2, isTop) { - var expansions = []; - var m = balanced("{", "}", str2); - if (!m) return [str2]; - var pre = m.pre; - var post = m.post.length ? expand(m.post, false) : [""]; + function expand_(str2, isTop) { + const expansions = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) + return [str2]; + const pre = m.pre; + const post = m.post.length ? expand_(m.post, false) : [""]; if (/\$$/.test(m.pre)) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + "{" + m.body + "}" + post[k]; + for (let k = 0; k < post.length; k++) { + const expansion = pre + "{" + m.body + "}" + post[k]; expansions.push(expansion); } } else { - var isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); - var isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); - var isSequence = isNumericSequence || isAlphaSequence; - var isOptions = m.body.indexOf(",") >= 0; + const isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); + const isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); + const isSequence = isNumericSequence || isAlphaSequence; + const isOptions = m.body.indexOf(",") >= 0; if (!isSequence && !isOptions) { if (m.post.match(/,(?!,).*\}/)) { str2 = m.pre + "{" + m.body + escClose + m.post; - return expand(str2); + return expand_(str2); } return [str2]; } - var n; + let n; if (isSequence) { n = m.body.split(/\.\./); } else { n = parseCommaParts(m.body); - if (n.length === 1) { - n = expand(n[0], false).map(embrace); + if (n.length === 1 && n[0] !== void 0) { + n = expand_(n[0], false).map(embrace); if (n.length === 1) { - return post.map(function(p) { - return m.pre + n[0] + p; - }); + return post.map((p) => m.pre + n[0] + p); } } } - var N; - if (isSequence) { - var x = numeric(n[0]); - var y = numeric(n[1]); - var width = Math.max(n[0].length, n[1].length); - var incr = n.length == 3 ? Math.abs(numeric(n[2])) : 1; - var test = lte; - var reverse = y < x; + let N; + if (isSequence && n[0] !== void 0 && n[1] !== void 0) { + const x = numeric(n[0]); + const y = numeric(n[1]); + const width = Math.max(n[0].length, n[1].length); + let incr = n.length === 3 && n[2] !== void 0 ? Math.abs(numeric(n[2])) : 1; + let test = lte; + const reverse = y < x; if (reverse) { incr *= -1; test = gte5; } - var pad = n.some(isPadded); + const pad = n.some(isPadded); N = []; - for (var i = x; test(i, y); i += incr) { - var c; + for (let i = x; test(i, y); i += incr) { + let c; if (isAlphaSequence) { c = String.fromCharCode(i); - if (c === "\\") + if (c === "\\") { c = ""; + } } else { c = String(i); if (pad) { - var need = width - c.length; + const need = width - c.length; if (need > 0) { - var z = new Array(need + 1).join("0"); - if (i < 0) + const z = new Array(need + 1).join("0"); + if (i < 0) { c = "-" + z + c.slice(1); - else + } else { c = z + c; + } } } } @@ -85826,15 +85904,16 @@ var require_brace_expansion2 = __commonJS({ } } else { N = []; - for (var j = 0; j < n.length; j++) { - N.push.apply(N, expand(n[j], false)); + for (let j = 0; j < n.length; j++) { + N.push.apply(N, expand_(n[j], false)); } } - for (var j = 0; j < N.length; j++) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + N[j] + post[k]; - if (!isTop || isSequence || expansion) + for (let j = 0; j < N.length; j++) { + for (let k = 0; k < post.length; k++) { + const expansion = pre + N[j] + post[k]; + if (!isTop || isSequence || expansion) { expansions.push(expansion); + } } } } @@ -85843,9 +85922,9 @@ var require_brace_expansion2 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js var require_assert_valid_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.assertValidPattern = void 0; @@ -85862,9 +85941,9 @@ var require_assert_valid_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js var require_brace_expressions = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.parseClass = void 0; @@ -85979,22 +86058,25 @@ var require_brace_expressions = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js var require_unescape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = void 0; - var unescape = (s, { windowsPathsNoEscape = false } = {}) => { - return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + var unescape = (s, { windowsPathsNoEscape = false, magicalBraces = true } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + } + return windowsPathsNoEscape ? s.replace(/\[([^\/\\{}])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\{}])\]/g, "$1$2").replace(/\\([^\/{}])/g, "$1"); }; exports2.unescape = unescape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js var require_ast = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.AST = void 0; @@ -86350,7 +86432,7 @@ var require_ast = __commonJS({ if (this.#root === this) this.#fillNegs(); if (!this.type) { - const noEmpty = this.isStart() && this.isEnd(); + const noEmpty = this.isStart() && this.isEnd() && !this.#parts.some((s) => typeof s !== "string"); const src = this.#parts.map((p) => { const [re, _2, hasMagic, uflag] = typeof p === "string" ? _AST.#parseGlob(p, this.#hasMagic, noEmpty) : p.toRegExpSource(allowDot); this.#hasMagic = this.#hasMagic || hasMagic; @@ -86460,10 +86542,7 @@ var require_ast = __commonJS({ } } if (c === "*") { - if (noEmpty && glob2 === "*") - re += starNoEmpty; - else - re += star; + re += noEmpty && glob2 === "*" ? starNoEmpty : star; hasMagic = true; continue; } @@ -86481,29 +86560,29 @@ var require_ast = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js var require_escape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.escape = void 0; - var escape = (s, { windowsPathsNoEscape = false } = {}) => { + var escape = (s, { windowsPathsNoEscape = false, magicalBraces = false } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/[?*()[\]{}]/g, "[$&]") : s.replace(/[?*()[\]\\{}]/g, "\\$&"); + } return windowsPathsNoEscape ? s.replace(/[?*()[\]]/g, "[$&]") : s.replace(/[?*()[\]\\]/g, "\\$&"); }; exports2.escape = escape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js -var require_commonjs13 = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js"(exports2) { +// node_modules/glob/node_modules/minimatch/dist/commonjs/index.js +var require_commonjs15 = __commonJS({ + "node_modules/glob/node_modules/minimatch/dist/commonjs/index.js"(exports2) { "use strict"; - var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { - return mod && mod.__esModule ? mod : { "default": mod }; - }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = exports2.escape = exports2.AST = exports2.Minimatch = exports2.match = exports2.makeRe = exports2.braceExpand = exports2.defaults = exports2.filter = exports2.GLOBSTAR = exports2.sep = exports2.minimatch = void 0; - var brace_expansion_1 = __importDefault4(require_brace_expansion2()); + var brace_expansion_1 = require_commonjs14(); var assert_valid_pattern_js_1 = require_assert_valid_pattern(); var ast_js_1 = require_ast(); var escape_js_1 = require_escape(); @@ -86626,7 +86705,7 @@ var require_commonjs13 = __commonJS({ if (options.nobrace || !/\{(?:(?!\{).)*\}/.test(pattern)) { return [pattern]; } - return (0, brace_expansion_1.default)(pattern); + return (0, brace_expansion_1.expand)(pattern); }; exports2.braceExpand = braceExpand; exports2.minimatch.braceExpand = exports2.braceExpand; @@ -87150,16 +87229,27 @@ var require_commonjs13 = __commonJS({ pp[i] = twoStar; } } else if (next === void 0) { - pp[i - 1] = prev + "(?:\\/|" + twoStar + ")?"; + pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + ")?"; } else if (next !== exports2.GLOBSTAR) { pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + "\\/)" + next; pp[i + 1] = exports2.GLOBSTAR; } }); - return pp.filter((p) => p !== exports2.GLOBSTAR).join("/"); + const filtered = pp.filter((p) => p !== exports2.GLOBSTAR); + if (this.partial && filtered.length >= 1) { + const prefixes = []; + for (let i = 1; i <= filtered.length; i++) { + prefixes.push(filtered.slice(0, i).join("/")); + } + return "(?:" + prefixes.join("|") + ")"; + } + return filtered.join("/"); }).join("|"); const [open, close] = set2.length > 1 ? ["(?:", ")"] : ["", ""]; re = "^" + open + re + close + "$"; + if (this.partial) { + re = "^(?:\\/|" + open + re.slice(1, -1) + close + ")$"; + } if (this.negate) re = "^(?!" + re + ").+$"; try { @@ -87246,9 +87336,9 @@ var require_commonjs13 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js -var require_commonjs14 = __commonJS({ - "node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js"(exports2) { +// node_modules/lru-cache/dist/commonjs/index.js +var require_commonjs16 = __commonJS({ + "node_modules/lru-cache/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.LRUCache = void 0; @@ -87337,6 +87427,7 @@ var require_commonjs14 = __commonJS({ #max; #maxSize; #dispose; + #onInsert; #disposeAfter; #fetchMethod; #memoMethod; @@ -87418,6 +87509,7 @@ var require_commonjs14 = __commonJS({ #hasDispose; #hasFetchMethod; #hasDisposeAfter; + #hasOnInsert; /** * Do not call this method unless you need to inspect the * inner workings of the cache. If anything returned by this @@ -87494,6 +87586,12 @@ var require_commonjs14 = __commonJS({ get dispose() { return this.#dispose; } + /** + * {@link LRUCache.OptionsBase.onInsert} (read-only) + */ + get onInsert() { + return this.#onInsert; + } /** * {@link LRUCache.OptionsBase.disposeAfter} (read-only) */ @@ -87501,7 +87599,7 @@ var require_commonjs14 = __commonJS({ return this.#disposeAfter; } constructor(options) { - const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; + const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, onInsert, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; if (max !== 0 && !isPosInt(max)) { throw new TypeError("max option must be a nonnegative integer"); } @@ -87543,6 +87641,9 @@ var require_commonjs14 = __commonJS({ if (typeof dispose === "function") { this.#dispose = dispose; } + if (typeof onInsert === "function") { + this.#onInsert = onInsert; + } if (typeof disposeAfter === "function") { this.#disposeAfter = disposeAfter; this.#disposed = []; @@ -87551,6 +87652,7 @@ var require_commonjs14 = __commonJS({ this.#disposed = void 0; } this.#hasDispose = !!this.#dispose; + this.#hasOnInsert = !!this.#onInsert; this.#hasDisposeAfter = !!this.#disposeAfter; this.noDisposeOnSet = !!noDisposeOnSet; this.noUpdateTTL = !!noUpdateTTL; @@ -87953,7 +88055,7 @@ var require_commonjs14 = __commonJS({ } /** * Return an array of [key, {@link LRUCache.Entry}] tuples which can be - * passed to {@link LRLUCache#load}. + * passed to {@link LRUCache#load}. * * The `start` fields are calculated relative to a portable `Date.now()` * timestamp, even if `performance.now()` is available. @@ -88064,6 +88166,9 @@ var require_commonjs14 = __commonJS({ if (status) status.set = "add"; noUpdateTTL = false; + if (this.#hasOnInsert) { + this.#onInsert?.(v, k, "add"); + } } else { this.#moveToTail(index); const oldVal = this.#valList[index]; @@ -88099,6 +88204,9 @@ var require_commonjs14 = __commonJS({ } else if (status) { status.set = "update"; } + if (this.#hasOnInsert) { + this.onInsert?.(v, k, v === oldVal ? "update" : "replace"); + } } if (ttl !== 0 && !this.#ttls) { this.#initializeTTLTracking(); @@ -88624,7 +88732,7 @@ var require_commonjs14 = __commonJS({ }); // node_modules/minipass/dist/commonjs/index.js -var require_commonjs15 = __commonJS({ +var require_commonjs17 = __commonJS({ "node_modules/minipass/dist/commonjs/index.js"(exports2) { "use strict"; var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { @@ -89515,9 +89623,9 @@ var require_commonjs15 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js -var require_commonjs16 = __commonJS({ - "node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js"(exports2) { +// node_modules/path-scurry/dist/commonjs/index.js +var require_commonjs18 = __commonJS({ + "node_modules/path-scurry/dist/commonjs/index.js"(exports2) { "use strict"; var __createBinding4 = exports2 && exports2.__createBinding || (Object.create ? (function(o, m, k, k2) { if (k2 === void 0) k2 = k; @@ -89548,14 +89656,14 @@ var require_commonjs16 = __commonJS({ }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.PathScurry = exports2.Path = exports2.PathScurryDarwin = exports2.PathScurryPosix = exports2.PathScurryWin32 = exports2.PathScurryBase = exports2.PathPosix = exports2.PathWin32 = exports2.PathBase = exports2.ChildrenCache = exports2.ResolveCache = void 0; - var lru_cache_1 = require_commonjs14(); + var lru_cache_1 = require_commonjs16(); var node_path_1 = require("node:path"); var node_url_1 = require("node:url"); var fs_1 = require("fs"); var actualFS = __importStar4(require("node:fs")); var realpathSync = fs_1.realpathSync.native; var promises_1 = require("node:fs/promises"); - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var defaultFS = { lstatSync: fs_1.lstatSync, readdir: fs_1.readdir, @@ -89770,6 +89878,8 @@ var require_commonjs16 = __commonJS({ /** * Deprecated alias for Dirent['parentPath'] Somewhat counterintuitively, * this property refers to the *parent* path, not the path object itself. + * + * @deprecated */ get path() { return this.parentPath; @@ -91289,13 +91399,13 @@ var require_commonjs16 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js +// node_modules/glob/dist/commonjs/pattern.js var require_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js"(exports2) { + "node_modules/glob/dist/commonjs/pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Pattern = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var isPatternList = (pl) => pl.length >= 1; var isGlobList = (gl) => gl.length >= 1; var Pattern = class _Pattern { @@ -91463,13 +91573,13 @@ var require_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js +// node_modules/glob/dist/commonjs/ignore.js var require_ignore = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js"(exports2) { + "node_modules/glob/dist/commonjs/ignore.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Ignore = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var pattern_js_1 = require_pattern(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; var Ignore = class { @@ -91560,13 +91670,13 @@ var require_ignore = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js +// node_modules/glob/dist/commonjs/processor.js var require_processor = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js"(exports2) { + "node_modules/glob/dist/commonjs/processor.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Processor = exports2.SubWalks = exports2.MatchRecord = exports2.HasWalkedCache = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var HasWalkedCache = class _HasWalkedCache { store; constructor(store = /* @__PURE__ */ new Map()) { @@ -91793,13 +91903,13 @@ var require_processor = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js +// node_modules/glob/dist/commonjs/walker.js var require_walker = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js"(exports2) { + "node_modules/glob/dist/commonjs/walker.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.GlobStream = exports2.GlobWalker = exports2.GlobUtil = void 0; - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var ignore_js_1 = require_ignore(); var processor_js_1 = require_processor(); var makeIgnore = (ignore, opts) => typeof ignore === "string" ? new ignore_js_1.Ignore([ignore], opts) : Array.isArray(ignore) ? new ignore_js_1.Ignore(ignore, opts) : ignore; @@ -92133,15 +92243,15 @@ var require_walker = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js +// node_modules/glob/dist/commonjs/glob.js var require_glob = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js"(exports2) { + "node_modules/glob/dist/commonjs/glob.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Glob = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var node_url_1 = require("node:url"); - var path_scurry_1 = require_commonjs16(); + var path_scurry_1 = require_commonjs18(); var pattern_js_1 = require_pattern(); var walker_js_1 = require_walker(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; @@ -92346,13 +92456,13 @@ var require_glob = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js +// node_modules/glob/dist/commonjs/has-magic.js var require_has_magic = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js"(exports2) { + "node_modules/glob/dist/commonjs/has-magic.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.hasMagic = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var hasMagic = (pattern, options = {}) => { if (!Array.isArray(pattern)) { pattern = [pattern]; @@ -92367,9 +92477,9 @@ var require_has_magic = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js -var require_commonjs17 = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js"(exports2) { +// node_modules/glob/dist/commonjs/index.js +var require_commonjs19 = __commonJS({ + "node_modules/glob/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.glob = exports2.sync = exports2.iterate = exports2.iterateSync = exports2.stream = exports2.streamSync = exports2.Ignore = exports2.hasMagic = exports2.Glob = exports2.unescape = exports2.escape = void 0; @@ -92378,10 +92488,10 @@ var require_commonjs17 = __commonJS({ exports2.globSync = globSync; exports2.globIterateSync = globIterateSync; exports2.globIterate = globIterate; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var glob_js_1 = require_glob(); var has_magic_js_1 = require_has_magic(); - var minimatch_2 = require_commonjs13(); + var minimatch_2 = require_commonjs15(); Object.defineProperty(exports2, "escape", { enumerable: true, get: function() { return minimatch_2.escape; } }); @@ -92458,7 +92568,7 @@ var require_file3 = __commonJS({ var difference = require_difference(); var union = require_union(); var isPlainObject = require_isPlainObject(); - var glob2 = require_commonjs17(); + var glob2 = require_commonjs19(); var file = module2.exports = {}; var pathSeparatorRe = /[\/\\]/g; var processPatterns = function(patterns, fn) { @@ -105273,7 +105383,7 @@ var require_concat_map = __commonJS({ }); // node_modules/brace-expansion/index.js -var require_brace_expansion3 = __commonJS({ +var require_brace_expansion2 = __commonJS({ "node_modules/brace-expansion/index.js"(exports2, module2) { var concatMap = require_concat_map(); var balanced = require_balanced_match(); @@ -105431,7 +105541,7 @@ var require_minimatch2 = __commonJS({ }; minimatch.sep = path2.sep; var GLOBSTAR = minimatch.GLOBSTAR = Minimatch.GLOBSTAR = {}; - var expand = require_brace_expansion3(); + var expand = require_brace_expansion2(); var plTypes = { "!": { open: "(?:(?!(?:", close: "))[^/]*?)" }, "?": { open: "(?:", close: ")?" }, diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 6db9c67bd..667acaa15 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -27694,14 +27694,14 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", - eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + eslint: "^8.57.1", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27725,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } diff --git a/package-lock.json b/package-lock.json index 1a2335428..80ef7cb61 100644 --- a/package-lock.json +++ b/package-lock.json @@ -59,7 +59,7 @@ "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - "glob": "^11.0.3", + "glob": "^11.1.0", "nock": "^14.0.10", "sinon": "^21.0.0", "typescript": "^5.9.3" @@ -1820,7 +1820,6 @@ "resolved": "https://registry.npmjs.org/@octokit/core/-/core-7.0.6.tgz", "integrity": "sha512-DhGl4xMVFGVIyMwswXeyzdL4uXD5OGILGX5N8Y+f6W7LhC1Ze2poSNrkF/fedpVDHEEZ+PHFW0vL14I+mm8K3Q==", "license": "MIT", - "peer": true, "dependencies": { "@octokit/auth-token": "^6.0.0", "@octokit/graphql": "^9.0.3", @@ -1991,7 +1990,6 @@ "resolved": "https://registry.npmjs.org/@octokit/core/-/core-5.2.2.tgz", "integrity": "sha512-/g2d4sW9nUDJOMz3mabVQvOGhVa4e/BN/Um7yca9Bb2XTzPPnfTWHWQg+IsEYO7M3Vx+EXvaM/I2pJWIMun1bg==", "license": "MIT", - "peer": true, "dependencies": { "@octokit/auth-token": "^4.0.0", "@octokit/graphql": "^7.1.0", @@ -2400,16 +2398,6 @@ "node": ">=8.0.0" } }, - "node_modules/@pkgjs/parseargs": { - "version": "0.11.0", - "resolved": "https://registry.npmjs.org/@pkgjs/parseargs/-/parseargs-0.11.0.tgz", - "integrity": "sha512-+1VkjdD0QBLPodGrJUeqarH8VAIvQODIbwh9XpP5Syisf7YoQgsJKPNFoqqLQlu+VQ/tVSshMR6loPMn8U+dPg==", - "license": "MIT", - "optional": true, - "engines": { - "node": ">=14" - } - }, "node_modules/@pkgr/core": { "version": "0.1.1", "resolved": "https://registry.npmjs.org/@pkgr/core/-/core-0.1.1.tgz", @@ -2954,7 +2942,6 @@ "integrity": "sha512-tK3GPFWbirvNgsNKto+UmB/cRtn6TZfyw0D6IKrW55n6Vbs7KJoZtI//kpTKzE/DUmmnAFD8/Ca46s7Obs92/w==", "dev": true, "license": "MIT", - "peer": true, "dependencies": { "@typescript-eslint/scope-manager": "8.46.4", "@typescript-eslint/types": "8.46.4", @@ -3515,93 +3502,6 @@ "node": ">=18" } }, - "node_modules/@vercel/nft/node_modules/brace-expansion": { - "version": "2.0.2", - "resolved": "https://registry.npmjs.org/brace-expansion/-/brace-expansion-2.0.2.tgz", - "integrity": "sha512-Jt0vHyM+jmUBqojB7E1NIYadt0vI0Qxjxd2TErW94wDz+E2LAm5vKMXXwg6ZZBTHPuUlDgQHKXvjGBdfcF1ZDQ==", - "dev": true, - "license": "MIT", - "dependencies": { - "balanced-match": "^1.0.0" - } - }, - "node_modules/@vercel/nft/node_modules/glob": { - "version": "10.4.5", - "resolved": "https://registry.npmjs.org/glob/-/glob-10.4.5.tgz", - "integrity": "sha512-7Bv8RF0k6xjo7d4A/PxYLbUCfb6c+Vpd2/mB2yRDlew7Jb5hEXiCD9ibfO7wpk8i4sevK6DFny9h7EYbM3/sHg==", - "dev": true, - "license": "ISC", - "dependencies": { - "foreground-child": "^3.1.0", - "jackspeak": "^3.1.2", - "minimatch": "^9.0.4", - "minipass": "^7.1.2", - "package-json-from-dist": "^1.0.0", - "path-scurry": "^1.11.1" - }, - "bin": { - "glob": "dist/esm/bin.mjs" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, - "node_modules/@vercel/nft/node_modules/jackspeak": { - "version": "3.4.3", - "resolved": "https://registry.npmjs.org/jackspeak/-/jackspeak-3.4.3.tgz", - "integrity": "sha512-OGlZQpz2yfahA/Rd1Y8Cd9SIEsqvXkLVoSw/cgwhnhFMDbsQFeZYoJJ7bIZBS9BcamUW96asq/npPWugM+RQBw==", - "dev": true, - "license": "BlueOak-1.0.0", - "dependencies": { - "@isaacs/cliui": "^8.0.2" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - }, - "optionalDependencies": { - "@pkgjs/parseargs": "^0.11.0" - } - }, - "node_modules/@vercel/nft/node_modules/lru-cache": { - "version": "10.4.3", - "resolved": "https://registry.npmjs.org/lru-cache/-/lru-cache-10.4.3.tgz", - "integrity": "sha512-JNAzZcXrCt42VGLuYz0zfAzDfAvJWW6AfYlDBQyDV5DClI2m5sAmK+OIO7s59XfsRsWHp02jAJrRadPRGTt6SQ==", - "dev": true, - "license": "ISC" - }, - "node_modules/@vercel/nft/node_modules/minimatch": { - "version": "9.0.5", - "resolved": "https://registry.npmjs.org/minimatch/-/minimatch-9.0.5.tgz", - "integrity": "sha512-G6T0ZX48xgozx7587koeX9Ys2NYy6Gmv//P89sEte9V9whIapMNF4idKxnW2QtCcLiTWlb/wfCabAtAFWhhBow==", - "dev": true, - "license": "ISC", - "dependencies": { - "brace-expansion": "^2.0.1" - }, - "engines": { - "node": ">=16 || 14 >=14.17" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, - "node_modules/@vercel/nft/node_modules/path-scurry": { - "version": "1.11.1", - "resolved": "https://registry.npmjs.org/path-scurry/-/path-scurry-1.11.1.tgz", - "integrity": "sha512-Xa4Nw17FS9ApQFJ9umLiJS4orGjm7ZzwUrwamcGQuHSzDyth9boKDaycYdDcZDuqYATXw4HFXgaqWTctW/v1HA==", - "dev": true, - "license": "BlueOak-1.0.0", - "dependencies": { - "lru-cache": "^10.2.0", - "minipass": "^5.0.0 || ^6.0.2 || ^7.0.0" - }, - "engines": { - "node": ">=16 || 14 >=14.18" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, "node_modules/@vercel/nft/node_modules/picomatch": { "version": "4.0.2", "resolved": "https://registry.npmjs.org/picomatch/-/picomatch-4.0.2.tgz", @@ -3638,7 +3538,6 @@ "resolved": "https://registry.npmjs.org/acorn/-/acorn-8.15.0.tgz", "integrity": "sha512-NZyJarBfL7nWwIq+FDL6Zp/yHEhePMNnnJ0y3qfieCrmNvYct8uvtiV41UvlSe6apAfk0fY1FbWx+NwfmpvtTg==", "dev": true, - "peer": true, "bin": { "acorn": "bin/acorn" }, @@ -3751,35 +3650,6 @@ "node": ">= 14" } }, - "node_modules/archiver-utils/node_modules/brace-expansion": { - "version": "2.0.2", - "resolved": "https://registry.npmjs.org/brace-expansion/-/brace-expansion-2.0.2.tgz", - "integrity": "sha512-Jt0vHyM+jmUBqojB7E1NIYadt0vI0Qxjxd2TErW94wDz+E2LAm5vKMXXwg6ZZBTHPuUlDgQHKXvjGBdfcF1ZDQ==", - "license": "MIT", - "dependencies": { - "balanced-match": "^1.0.0" - } - }, - "node_modules/archiver-utils/node_modules/glob": { - "version": "10.4.5", - "resolved": "https://registry.npmjs.org/glob/-/glob-10.4.5.tgz", - "integrity": "sha512-7Bv8RF0k6xjo7d4A/PxYLbUCfb6c+Vpd2/mB2yRDlew7Jb5hEXiCD9ibfO7wpk8i4sevK6DFny9h7EYbM3/sHg==", - "license": "ISC", - "dependencies": { - "foreground-child": "^3.1.0", - "jackspeak": "^3.1.2", - "minimatch": "^9.0.4", - "minipass": "^7.1.2", - "package-json-from-dist": "^1.0.0", - "path-scurry": "^1.11.1" - }, - "bin": { - "glob": "dist/esm/bin.mjs" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, "node_modules/archiver-utils/node_modules/is-stream": { "version": "2.0.1", "resolved": "https://registry.npmjs.org/is-stream/-/is-stream-2.0.1.tgz", @@ -3792,58 +3662,6 @@ "url": "https://github.com/sponsors/sindresorhus" } }, - "node_modules/archiver-utils/node_modules/jackspeak": { - "version": "3.4.3", - "resolved": "https://registry.npmjs.org/jackspeak/-/jackspeak-3.4.3.tgz", - "integrity": "sha512-OGlZQpz2yfahA/Rd1Y8Cd9SIEsqvXkLVoSw/cgwhnhFMDbsQFeZYoJJ7bIZBS9BcamUW96asq/npPWugM+RQBw==", - "license": "BlueOak-1.0.0", - "dependencies": { - "@isaacs/cliui": "^8.0.2" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - }, - "optionalDependencies": { - "@pkgjs/parseargs": "^0.11.0" - } - }, - "node_modules/archiver-utils/node_modules/lru-cache": { - "version": "10.4.3", - "resolved": "https://registry.npmjs.org/lru-cache/-/lru-cache-10.4.3.tgz", - "integrity": "sha512-JNAzZcXrCt42VGLuYz0zfAzDfAvJWW6AfYlDBQyDV5DClI2m5sAmK+OIO7s59XfsRsWHp02jAJrRadPRGTt6SQ==", - "license": "ISC" - }, - "node_modules/archiver-utils/node_modules/minimatch": { - "version": "9.0.5", - "resolved": "https://registry.npmjs.org/minimatch/-/minimatch-9.0.5.tgz", - "integrity": "sha512-G6T0ZX48xgozx7587koeX9Ys2NYy6Gmv//P89sEte9V9whIapMNF4idKxnW2QtCcLiTWlb/wfCabAtAFWhhBow==", - "license": "ISC", - "dependencies": { - "brace-expansion": "^2.0.1" - }, - "engines": { - "node": ">=16 || 14 >=14.17" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, - "node_modules/archiver-utils/node_modules/path-scurry": { - "version": "1.11.1", - "resolved": "https://registry.npmjs.org/path-scurry/-/path-scurry-1.11.1.tgz", - "integrity": "sha512-Xa4Nw17FS9ApQFJ9umLiJS4orGjm7ZzwUrwamcGQuHSzDyth9boKDaycYdDcZDuqYATXw4HFXgaqWTctW/v1HA==", - "license": "BlueOak-1.0.0", - "dependencies": { - "lru-cache": "^10.2.0", - "minipass": "^5.0.0 || ^6.0.2 || ^7.0.0" - }, - "engines": { - "node": ">=16 || 14 >=14.18" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, "node_modules/are-docs-informative": { "version": "0.0.2", "resolved": "https://registry.npmjs.org/are-docs-informative/-/are-docs-informative-0.0.2.tgz", @@ -4295,7 +4113,6 @@ } ], "license": "MIT", - "peer": true, "dependencies": { "caniuse-lite": "^1.0.30001669", "electron-to-chromium": "^1.5.41", @@ -5149,7 +4966,6 @@ "resolved": "https://registry.npmjs.org/eslint/-/eslint-8.57.1.tgz", "integrity": "sha512-ypowyDxpVSYpkXr9WPv2PAZCtNip1Mv5KTW0SCurXv/9iOpcrH9PaqUElksqEB6pChqHGDRCFTyrZlGhnLNGiA==", "dev": true, - "peer": true, "dependencies": { "@eslint-community/eslint-utils": "^4.2.0", "@eslint-community/regexpp": "^4.6.1", @@ -5204,7 +5020,6 @@ "version": "8.3.0", "dev": true, "license": "MIT", - "peer": true, "bin": { "eslint-config-prettier": "bin/cli.js" }, @@ -5442,7 +5257,6 @@ "resolved": "https://registry.npmjs.org/eslint-plugin-import/-/eslint-plugin-import-2.29.1.tgz", "integrity": "sha512-BbPC0cuExzhiMo4Ff1BTVwHpjjv28C5R+btTOGaCRC7UEz801up0JadwkeSk5Ued6TG34uaczuVuH6qyy5YUxw==", "dev": true, - "peer": true, "dependencies": { "array-includes": "^3.1.7", "array.prototype.findlastindex": "^1.2.3", @@ -6206,11 +6020,6 @@ "node": ">= 0.12" } }, - "node_modules/fs.realpath": { - "version": "1.0.0", - "dev": true, - "license": "ISC" - }, "node_modules/function-bind": { "version": "1.1.2", "license": "MIT", @@ -6359,15 +6168,15 @@ } }, "node_modules/glob": { - "version": "11.0.3", - "resolved": "https://registry.npmjs.org/glob/-/glob-11.0.3.tgz", - "integrity": "sha512-2Nim7dha1KVkaiF4q6Dj+ngPPMdfvLJEOpZk/jKiUAkqKebpGAWQXAq9z1xu9HKu5lWfqw/FASuccEjyznjPaA==", + "version": "11.1.0", + "resolved": "https://registry.npmjs.org/glob/-/glob-11.1.0.tgz", + "integrity": "sha512-vuNwKSaKiqm7g0THUBu2x7ckSs3XJLXE+2ssL7/MfTGPLLcrJQ/4Uq1CjPTtO5cCIiRxqvN6Twy1qOwhL0Xjcw==", "dev": true, - "license": "ISC", + "license": "BlueOak-1.0.0", "dependencies": { "foreground-child": "^3.3.1", "jackspeak": "^4.1.1", - "minimatch": "^10.0.3", + "minimatch": "^10.1.1", "minipass": "^7.1.2", "package-json-from-dist": "^1.0.0", "path-scurry": "^2.0.0" @@ -6394,11 +6203,11 @@ } }, "node_modules/glob/node_modules/minimatch": { - "version": "10.0.3", - "resolved": "https://registry.npmjs.org/minimatch/-/minimatch-10.0.3.tgz", - "integrity": "sha512-IPZ167aShDZZUMdRk66cyQAW3qr0WzbHkPdMYa8bzZhlHhO3jALbKdxcaak7W9FfT2rZNpQuUu4Od7ILEpXSaw==", + "version": "10.1.1", + "resolved": "https://registry.npmjs.org/minimatch/-/minimatch-10.1.1.tgz", + "integrity": "sha512-enIvLvRAFZYXJzkCYG5RKmPfrFArdLv+R+lbQ53BmIMLIry74bjKzX6iHAm8WYamJkhSSEabrWN5D97XnKObjQ==", "dev": true, - "license": "ISC", + "license": "BlueOak-1.0.0", "dependencies": { "@isaacs/brace-expansion": "^5.0.0" }, @@ -6710,15 +6519,6 @@ "url": "https://github.com/sponsors/sindresorhus" } }, - "node_modules/inflight": { - "version": "1.0.6", - "dev": true, - "license": "ISC", - "dependencies": { - "once": "^1.3.0", - "wrappy": "1" - } - }, "node_modules/inherits": { "version": "2.0.3", "license": "ISC" @@ -7744,7 +7544,6 @@ "resolved": "https://registry.npmjs.org/@octokit/core/-/core-7.0.6.tgz", "integrity": "sha512-DhGl4xMVFGVIyMwswXeyzdL4uXD5OGILGX5N8Y+f6W7LhC1Ze2poSNrkF/fedpVDHEEZ+PHFW0vL14I+mm8K3Q==", "license": "MIT", - "peer": true, "dependencies": { "@octokit/auth-token": "^6.0.0", "@octokit/graphql": "^9.0.3", @@ -7961,14 +7760,6 @@ "dev": true, "license": "MIT" }, - "node_modules/path-is-absolute": { - "version": "1.0.1", - "dev": true, - "license": "MIT", - "engines": { - "node": ">=0.10.0" - } - }, "node_modules/path-key": { "version": "3.1.1", "license": "MIT", @@ -8067,7 +7858,6 @@ "integrity": "sha512-G+YdqtITVZmOJje6QkXQWzl3fSfMxFwm1tjTyo9exhkmWSqC4Yhd1+lug++IlR2mvRVAxEDDWYkQdeSztajqgg==", "dev": true, "license": "MIT", - "peer": true, "bin": { "prettier": "bin/prettier.cjs" }, @@ -8324,25 +8114,6 @@ "url": "https://github.com/sponsors/isaacs" } }, - "node_modules/rimraf/node_modules/glob": { - "version": "7.2.0", - "dev": true, - "license": "ISC", - "dependencies": { - "fs.realpath": "^1.0.0", - "inflight": "^1.0.4", - "inherits": "2", - "minimatch": "^3.0.4", - "once": "^1.3.0", - "path-is-absolute": "^1.0.0" - }, - "engines": { - "node": "*" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, "node_modules/run-parallel": { "version": "1.2.0", "dev": true, @@ -9077,7 +8848,6 @@ "integrity": "sha512-M7BAV6Rlcy5u+m6oPhAPFgJTzAioX/6B0DxyvDlo9l8+T3nLKbrczg2WLUyzd45L8RqfUMyGPzekbMvX2Ldkwg==", "dev": true, "license": "MIT", - "peer": true, "engines": { "node": ">=12" }, @@ -9295,7 +9065,6 @@ "integrity": "sha512-jl1vZzPDinLr9eUt3J/t7V6FgNEw9QjvBPdysz9KfQDD41fQrC2Y4vKQdiaUpFT4bXlb1RHhLpp8wtm6M5TgSw==", "dev": true, "license": "Apache-2.0", - "peer": true, "bin": { "tsc": "bin/tsc", "tsserver": "bin/tsserver" @@ -9369,7 +9138,6 @@ "resolved": "https://registry.npmjs.org/@typescript-eslint/parser/-/parser-8.17.0.tgz", "integrity": "sha512-Drp39TXuUlD49F7ilHHCG7TTg8IkA+hxCuULdmzWYICxGXvDXmDmWEjJYZQYgf6l/TFfYNE167m7isnc3xlIEg==", "dev": true, - "peer": true, "dependencies": { "@typescript-eslint/scope-manager": "8.17.0", "@typescript-eslint/types": "8.17.0", diff --git a/package.json b/package.json index 5a3378cea..6eedf7f47 100644 --- a/package.json +++ b/package.json @@ -67,14 +67,14 @@ "@typescript-eslint/parser": "^8.41.0", "ava": "^6.4.1", "esbuild": "^0.27.0", - "eslint": "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - "glob": "^11.0.3", + "eslint": "^8.57.1", + "glob": "^11.1.0", "nock": "^14.0.10", "sinon": "^21.0.0", "typescript": "^5.9.3" @@ -98,6 +98,7 @@ "eslint-plugin-jsx-a11y": { "semver": ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + "glob": "^11.1.0" } } From 726a2a01b88c9c985274879d4c22c454458ed57a Mon Sep 17 00:00:00 2001 From: Kasper Svendsen Date: Tue, 18 Nov 2025 15:37:27 +0100 Subject: [PATCH 31/51] Overlay: Increase disk storage threshold to 20GB --- lib/analyze-action-post.js | 2 +- lib/analyze-action.js | 2 +- lib/autobuild-action.js | 2 +- lib/init-action-post.js | 2 +- lib/init-action.js | 2 +- lib/resolve-environment-action.js | 2 +- lib/setup-codeql-action.js | 2 +- lib/start-proxy-action-post.js | 2 +- lib/start-proxy-action.js | 2 +- lib/upload-lib.js | 2 +- lib/upload-sarif-action-post.js | 2 +- lib/upload-sarif-action.js | 2 +- src/config-utils.ts | 2 +- 13 files changed, 13 insertions(+), 13 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index 3dcda989f..de3b89b11 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -120102,7 +120102,7 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts -var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 11b32eac9..d80524b30 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -89370,7 +89370,7 @@ async function cachePrefix(codeql, language) { } // src/config-utils.ts -var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index 0a5626575..e9807b9d6 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -84416,7 +84416,7 @@ var GitHubFeatureFlags = class { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts -var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, diff --git a/lib/init-action-post.js b/lib/init-action-post.js index 6ab23eb5a..45f90f035 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -123766,7 +123766,7 @@ ${jsonContents}` var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts -var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, diff --git a/lib/init-action.js b/lib/init-action.js index 0630127df..ea9246bb5 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -86653,7 +86653,7 @@ async function cachePrefix(codeql, language) { } // src/config-utils.ts -var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; async function getSupportedLanguageMap(codeql, logger) { const resolveSupportedLanguagesUsingCli = await codeql.supportsFeature( diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index 3cca7431d..8bfc24ca0 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -84142,7 +84142,7 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts -var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index 0d0df74a4..ccd2157ad 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -84587,7 +84587,7 @@ var PACK_IDENTIFIER_PATTERN = (function() { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts -var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index 9377bcea4..c22306a3b 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -119508,7 +119508,7 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts -var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index 26d0fa13b..e54c456eb 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -100170,7 +100170,7 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts -var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, diff --git a/lib/upload-lib.js b/lib/upload-lib.js index be50f109a..3b951984b 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -87226,7 +87226,7 @@ ${jsonContents}` var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts -var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index af2299f7f..4c6f13802 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -119674,7 +119674,7 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts -var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index ea4b0e31b..b6eaf19a0 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -87307,7 +87307,7 @@ ${jsonContents}` var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts -var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15e3; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, diff --git a/src/config-utils.ts b/src/config-utils.ts index 6a1c40b13..f910e5d90 100644 --- a/src/config-utils.ts +++ b/src/config-utils.ts @@ -55,7 +55,7 @@ export * from "./config/db-config"; * analysis unless overlay analysis has been explicitly enabled via environment * variable. */ -const OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 15000; +const OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 20000; const OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1_000_000; From c9cb6f9c13e4f332e53ed0b3c512042839d798d0 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Tue, 18 Nov 2025 15:18:43 +0000 Subject: [PATCH 32/51] Update changelog for v4.31.4 --- CHANGELOG.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 4ccc60273..a3117206d 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -2,7 +2,7 @@ See the [releases page](https://github.com/github/codeql-action/releases) for the relevant changes to the CodeQL CLI and language packs. -## [UNRELEASED] +## 4.31.4 - 18 Nov 2025 No user facing changes. From fea250010cc47adedd75a7856be61ea11e54ab85 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Tue, 18 Nov 2025 16:14:11 +0000 Subject: [PATCH 33/51] Update changelog and version after v4.31.4 --- CHANGELOG.md | 4 ++++ package-lock.json | 4 ++-- package.json | 2 +- 3 files changed, 7 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index a3117206d..3ec876d4e 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -2,6 +2,10 @@ See the [releases page](https://github.com/github/codeql-action/releases) for the relevant changes to the CodeQL CLI and language packs. +## [UNRELEASED] + +No user facing changes. + ## 4.31.4 - 18 Nov 2025 No user facing changes. diff --git a/package-lock.json b/package-lock.json index 80ef7cb61..e7d1ab33b 100644 --- a/package-lock.json +++ b/package-lock.json @@ -1,12 +1,12 @@ { "name": "codeql", - "version": "4.31.4", + "version": "4.31.5", "lockfileVersion": 3, "requires": true, "packages": { "": { "name": "codeql", - "version": "4.31.4", + "version": "4.31.5", "license": "MIT", "dependencies": { "@actions/artifact": "^4.0.0", diff --git a/package.json b/package.json index 6eedf7f47..e318c8ebe 100644 --- a/package.json +++ b/package.json @@ -1,6 +1,6 @@ { "name": "codeql", - "version": "4.31.4", + "version": "4.31.5", "private": true, "description": "CodeQL action", "scripts": { From ce9b5264482ff71623ab959b094473a6717c35fa Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Tue, 18 Nov 2025 16:17:35 +0000 Subject: [PATCH 34/51] Rebuild --- lib/analyze-action-post.js | 2 +- lib/analyze-action.js | 2 +- lib/autobuild-action.js | 2 +- lib/init-action-post.js | 2 +- lib/init-action.js | 2 +- lib/resolve-environment-action.js | 2 +- lib/setup-codeql-action.js | 2 +- lib/start-proxy-action-post.js | 2 +- lib/start-proxy-action.js | 2 +- lib/upload-lib.js | 2 +- lib/upload-sarif-action-post.js | 2 +- lib/upload-sarif-action.js | 2 +- 12 files changed, 12 insertions(+), 12 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index 7dde941ee..ec0a2d876 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.4", + version: "4.31.5", private: true, description: "CodeQL action", scripts: { diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 242a05aca..ab65ec176 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.4", + version: "4.31.5", private: true, description: "CodeQL action", scripts: { diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index 3049d15a2..05344f9c9 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.4", + version: "4.31.5", private: true, description: "CodeQL action", scripts: { diff --git a/lib/init-action-post.js b/lib/init-action-post.js index 5dd89dcf6..6671e7f67 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.4", + version: "4.31.5", private: true, description: "CodeQL action", scripts: { diff --git a/lib/init-action.js b/lib/init-action.js index 4d68e5c9a..2e5384b4b 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.4", + version: "4.31.5", private: true, description: "CodeQL action", scripts: { diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index c3d54f680..bb774d967 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.4", + version: "4.31.5", private: true, description: "CodeQL action", scripts: { diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index 973e9c431..551338a38 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.4", + version: "4.31.5", private: true, description: "CodeQL action", scripts: { diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index 7e34a5e95..1669c4770 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.4", + version: "4.31.5", private: true, description: "CodeQL action", scripts: { diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index c0869dd96..7f991a6af 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -47285,7 +47285,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.4", + version: "4.31.5", private: true, description: "CodeQL action", scripts: { diff --git a/lib/upload-lib.js b/lib/upload-lib.js index ea6e2ca41..d6c9fed6b 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -28924,7 +28924,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.4", + version: "4.31.5", private: true, description: "CodeQL action", scripts: { diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index fce0c4f79..aa3effa71 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.4", + version: "4.31.5", private: true, description: "CodeQL action", scripts: { diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 667acaa15..9fd82027f 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.4", + version: "4.31.5", private: true, description: "CodeQL action", scripts: { From e24190a70c7ae613957f9b0205bd59708d3508dc Mon Sep 17 00:00:00 2001 From: Henry Mercer Date: Tue, 18 Nov 2025 18:02:59 +0000 Subject: [PATCH 35/51] Remove unused dependencies --- lib/analyze-action-post.js | 6 +- lib/analyze-action.js | 6 +- lib/autobuild-action.js | 6 +- lib/init-action-post.js | 6 +- lib/init-action.js | 6 +- lib/resolve-environment-action.js | 6 +- lib/setup-codeql-action.js | 6 +- lib/start-proxy-action-post.js | 6 +- lib/start-proxy-action.js | 6 +- lib/upload-lib.js | 6 +- lib/upload-sarif-action-post.js | 6 +- lib/upload-sarif-action.js | 6 +- package-lock.json | 575 +----------------------------- package.json | 6 +- 14 files changed, 36 insertions(+), 617 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index 7dde941ee..6da36a092 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -27662,7 +27662,6 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27672,7 +27671,6 @@ var require_package = __commonJS({ jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, @@ -27686,7 +27684,7 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", "@types/sinon": "^17.0.4", @@ -27694,13 +27692,13 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", + eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - eslint: "^8.57.1", glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 242a05aca..22666ed2a 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -27662,7 +27662,6 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27672,7 +27671,6 @@ var require_package = __commonJS({ jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, @@ -27686,7 +27684,7 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", "@types/sinon": "^17.0.4", @@ -27694,13 +27692,13 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", + eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - eslint: "^8.57.1", glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index 3049d15a2..19055f637 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -27662,7 +27662,6 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27672,7 +27671,6 @@ var require_package = __commonJS({ jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, @@ -27686,7 +27684,7 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", "@types/sinon": "^17.0.4", @@ -27694,13 +27692,13 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", + eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - eslint: "^8.57.1", glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", diff --git a/lib/init-action-post.js b/lib/init-action-post.js index 5dd89dcf6..5857024ff 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -27662,7 +27662,6 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27672,7 +27671,6 @@ var require_package = __commonJS({ jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, @@ -27686,7 +27684,7 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", "@types/sinon": "^17.0.4", @@ -27694,13 +27692,13 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", + eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - eslint: "^8.57.1", glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", diff --git a/lib/init-action.js b/lib/init-action.js index 4d68e5c9a..267a67e31 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -27662,7 +27662,6 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27672,7 +27671,6 @@ var require_package = __commonJS({ jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, @@ -27686,7 +27684,7 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", "@types/sinon": "^17.0.4", @@ -27694,13 +27692,13 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", + eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - eslint: "^8.57.1", glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index c3d54f680..d5769de3a 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -27662,7 +27662,6 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27672,7 +27671,6 @@ var require_package = __commonJS({ jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, @@ -27686,7 +27684,7 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", "@types/sinon": "^17.0.4", @@ -27694,13 +27692,13 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", + eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - eslint: "^8.57.1", glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index 973e9c431..b9577ce2a 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -27662,7 +27662,6 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27672,7 +27671,6 @@ var require_package = __commonJS({ jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, @@ -27686,7 +27684,7 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", "@types/sinon": "^17.0.4", @@ -27694,13 +27692,13 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", + eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - eslint: "^8.57.1", glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index 7e34a5e95..894cec283 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -27662,7 +27662,6 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27672,7 +27671,6 @@ var require_package = __commonJS({ jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, @@ -27686,7 +27684,7 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", "@types/sinon": "^17.0.4", @@ -27694,13 +27692,13 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", + eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - eslint: "^8.57.1", glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index c0869dd96..378fe7c8a 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -47320,7 +47320,6 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -47330,7 +47329,6 @@ var require_package = __commonJS({ jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, @@ -47344,7 +47342,7 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", "@types/sinon": "^17.0.4", @@ -47352,13 +47350,13 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", + eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - eslint: "^8.57.1", glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", diff --git a/lib/upload-lib.js b/lib/upload-lib.js index ea6e2ca41..a0881cc10 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -28959,7 +28959,6 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -28969,7 +28968,6 @@ var require_package = __commonJS({ jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, @@ -28983,7 +28981,7 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", "@types/sinon": "^17.0.4", @@ -28991,13 +28989,13 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", + eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - eslint: "^8.57.1", glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index fce0c4f79..55a3abf8c 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -27662,7 +27662,6 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27672,7 +27671,6 @@ var require_package = __commonJS({ jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, @@ -27686,7 +27684,7 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", "@types/sinon": "^17.0.4", @@ -27694,13 +27692,13 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", + eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - eslint: "^8.57.1", glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 667acaa15..25f9547cc 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -27662,7 +27662,6 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -27672,7 +27671,6 @@ var require_package = __commonJS({ jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, @@ -27686,7 +27684,7 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", "@types/sinon": "^17.0.4", @@ -27694,13 +27692,13 @@ var require_package = __commonJS({ "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", esbuild: "^0.27.0", + eslint: "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - eslint: "^8.57.1", glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", diff --git a/package-lock.json b/package-lock.json index 80ef7cb61..b8e9648ed 100644 --- a/package-lock.json +++ b/package-lock.json @@ -20,7 +20,6 @@ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", "archiver": "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -30,7 +29,6 @@ "jsonschema": "1.4.1", "long": "^5.3.2", "node-forge": "^1.3.1", - "octokit": "^5.0.5", "semver": "^7.7.3", "uuid": "^13.0.0" }, @@ -44,7 +42,7 @@ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", "@types/sinon": "^17.0.4", @@ -1578,7 +1576,6 @@ "version": "4.0.1", "resolved": "https://registry.npmjs.org/@isaacs/balanced-match/-/balanced-match-4.0.1.tgz", "integrity": "sha512-yzMTt9lEb8Gv7zRioUilSglI0c0smZ9k5D65677DLWLtWJaXIS3CqcGyUFByYKlnUj6TkjLVs54fBl6+TiGQDQ==", - "dev": true, "license": "MIT", "engines": { "node": "20 || >=22" @@ -1588,7 +1585,6 @@ "version": "5.0.0", "resolved": "https://registry.npmjs.org/@isaacs/brace-expansion/-/brace-expansion-5.0.0.tgz", "integrity": "sha512-ZT55BDLV0yv0RBm2czMiZ+SqCGO7AvmOM3G/w2xhVPH+te0aKgFjmBvGlL1dH+ql2tgGO3MVrbb3jCKyvpgnxA==", - "dev": true, "license": "MIT", "dependencies": { "@isaacs/balanced-match": "^4.0.1" @@ -1719,6 +1715,7 @@ "resolved": "https://registry.npmjs.org/@microsoft/eslint-formatter-sarif/-/eslint-formatter-sarif-3.1.0.tgz", "integrity": "sha512-/mn4UXziHzGXnKCg+r8HGgPy+w4RzpgdoqFuqaKOqUVBT5x2CygGefIrO4SusaY7t0C4gyIWMNu6YQT6Jw64Cw==", "dev": true, + "license": "MIT", "dependencies": { "eslint": "^8.9.0", "jschardet": "latest", @@ -1788,182 +1785,6 @@ "node": ">=12.4.0" } }, - "node_modules/@octokit/app": { - "version": "16.1.2", - "resolved": "https://registry.npmjs.org/@octokit/app/-/app-16.1.2.tgz", - "integrity": "sha512-8j7sEpUYVj18dxvh0KWj6W/l6uAiVRBl1JBDVRqH1VHKAO/G5eRVl4yEoYACjakWers1DjUkcCHyJNQK47JqyQ==", - "license": "MIT", - "dependencies": { - "@octokit/auth-app": "^8.1.2", - "@octokit/auth-unauthenticated": "^7.0.3", - "@octokit/core": "^7.0.6", - "@octokit/oauth-app": "^8.0.3", - "@octokit/plugin-paginate-rest": "^14.0.0", - "@octokit/types": "^16.0.0", - "@octokit/webhooks": "^14.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/app/node_modules/@octokit/auth-token": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/@octokit/auth-token/-/auth-token-6.0.0.tgz", - "integrity": "sha512-P4YJBPdPSpWTQ1NU4XYdvHvXJJDxM6YwpS0FZHRgP7YFkdVxsWcpWGy/NVqlAA7PcPCnMacXlRm1y2PFZRWL/w==", - "license": "MIT", - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/app/node_modules/@octokit/core": { - "version": "7.0.6", - "resolved": "https://registry.npmjs.org/@octokit/core/-/core-7.0.6.tgz", - "integrity": "sha512-DhGl4xMVFGVIyMwswXeyzdL4uXD5OGILGX5N8Y+f6W7LhC1Ze2poSNrkF/fedpVDHEEZ+PHFW0vL14I+mm8K3Q==", - "license": "MIT", - "dependencies": { - "@octokit/auth-token": "^6.0.0", - "@octokit/graphql": "^9.0.3", - "@octokit/request": "^10.0.6", - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0", - "before-after-hook": "^4.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/app/node_modules/@octokit/graphql": { - "version": "9.0.3", - "resolved": "https://registry.npmjs.org/@octokit/graphql/-/graphql-9.0.3.tgz", - "integrity": "sha512-grAEuupr/C1rALFnXTv6ZQhFuL1D8G5y8CN04RgrO4FIPMrtm+mcZzFG7dcBm+nq+1ppNixu+Jd78aeJOYxlGA==", - "license": "MIT", - "dependencies": { - "@octokit/request": "^10.0.6", - "@octokit/types": "^16.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/app/node_modules/@octokit/plugin-paginate-rest": { - "version": "14.0.0", - "resolved": "https://registry.npmjs.org/@octokit/plugin-paginate-rest/-/plugin-paginate-rest-14.0.0.tgz", - "integrity": "sha512-fNVRE7ufJiAA3XUrha2omTA39M6IXIc6GIZLvlbsm8QOQCYvpq/LkMNGyFlB1d8hTDzsAXa3OKtybdMAYsV/fw==", - "license": "MIT", - "dependencies": { - "@octokit/types": "^16.0.0" - }, - "engines": { - "node": ">= 20" - }, - "peerDependencies": { - "@octokit/core": ">=6" - } - }, - "node_modules/@octokit/app/node_modules/before-after-hook": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/before-after-hook/-/before-after-hook-4.0.0.tgz", - "integrity": "sha512-q6tR3RPqIB1pMiTRMFcZwuG5T8vwp+vUvEG0vuI6B+Rikh5BfPp2fQ82c925FOs+b0lcFQ8CFrL+KbilfZFhOQ==", - "license": "Apache-2.0" - }, - "node_modules/@octokit/app/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, - "node_modules/@octokit/auth-app": { - "version": "8.1.2", - "resolved": "https://registry.npmjs.org/@octokit/auth-app/-/auth-app-8.1.2.tgz", - "integrity": "sha512-db8VO0PqXxfzI6GdjtgEFHY9tzqUql5xMFXYA12juq8TeTgPAuiiP3zid4h50lwlIP457p5+56PnJOgd2GGBuw==", - "license": "MIT", - "dependencies": { - "@octokit/auth-oauth-app": "^9.0.3", - "@octokit/auth-oauth-user": "^6.0.2", - "@octokit/request": "^10.0.6", - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0", - "toad-cache": "^3.7.0", - "universal-github-app-jwt": "^2.2.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/auth-app/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, - "node_modules/@octokit/auth-oauth-app": { - "version": "9.0.3", - "resolved": "https://registry.npmjs.org/@octokit/auth-oauth-app/-/auth-oauth-app-9.0.3.tgz", - "integrity": "sha512-+yoFQquaF8OxJSxTb7rnytBIC2ZLbLqA/yb71I4ZXT9+Slw4TziV9j/kyGhUFRRTF2+7WlnIWsePZCWHs+OGjg==", - "license": "MIT", - "dependencies": { - "@octokit/auth-oauth-device": "^8.0.3", - "@octokit/auth-oauth-user": "^6.0.2", - "@octokit/request": "^10.0.6", - "@octokit/types": "^16.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/auth-oauth-app/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, - "node_modules/@octokit/auth-oauth-device": { - "version": "8.0.3", - "resolved": "https://registry.npmjs.org/@octokit/auth-oauth-device/-/auth-oauth-device-8.0.3.tgz", - "integrity": "sha512-zh2W0mKKMh/VWZhSqlaCzY7qFyrgd9oTWmTmHaXnHNeQRCZr/CXy2jCgHo4e4dJVTiuxP5dLa0YM5p5QVhJHbw==", - "license": "MIT", - "dependencies": { - "@octokit/oauth-methods": "^6.0.2", - "@octokit/request": "^10.0.6", - "@octokit/types": "^16.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/auth-oauth-device/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, - "node_modules/@octokit/auth-oauth-user": { - "version": "6.0.2", - "resolved": "https://registry.npmjs.org/@octokit/auth-oauth-user/-/auth-oauth-user-6.0.2.tgz", - "integrity": "sha512-qLoPPc6E6GJoz3XeDG/pnDhJpTkODTGG4kY0/Py154i/I003O9NazkrwJwRuzgCalhzyIeWQ+6MDvkUmKXjg/A==", - "license": "MIT", - "dependencies": { - "@octokit/auth-oauth-device": "^8.0.3", - "@octokit/oauth-methods": "^6.0.2", - "@octokit/request": "^10.0.6", - "@octokit/types": "^16.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/auth-oauth-user/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, "node_modules/@octokit/auth-token": { "version": "4.0.0", "resolved": "https://registry.npmjs.org/@octokit/auth-token/-/auth-token-4.0.0.tgz", @@ -1972,19 +1793,6 @@ "node": ">= 18" } }, - "node_modules/@octokit/auth-unauthenticated": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/@octokit/auth-unauthenticated/-/auth-unauthenticated-7.0.3.tgz", - "integrity": "sha512-8Jb1mtUdmBHL7lGmop9mU9ArMRUTRhg8vp0T1VtZ4yd9vEm3zcLwmjQkhNEduKawOOORie61xhtYIhTDN+ZQ3g==", - "license": "MIT", - "dependencies": { - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0" - }, - "engines": { - "node": ">= 20" - } - }, "node_modules/@octokit/core": { "version": "5.2.2", "resolved": "https://registry.npmjs.org/@octokit/core/-/core-5.2.2.tgz", @@ -2055,25 +1863,6 @@ "@octokit/openapi-types": "^24.2.0" } }, - "node_modules/@octokit/endpoint": { - "version": "11.0.2", - "resolved": "https://registry.npmjs.org/@octokit/endpoint/-/endpoint-11.0.2.tgz", - "integrity": "sha512-4zCpzP1fWc7QlqunZ5bSEjxc6yLAlRTnDwKtgXfcI/FxxGoqedDG8V2+xJ60bV2kODqcGB+nATdtap/XYq2NZQ==", - "license": "MIT", - "dependencies": { - "@octokit/types": "^16.0.0", - "universal-user-agent": "^7.0.2" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/endpoint/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, "node_modules/@octokit/graphql": { "version": "7.1.1", "resolved": "https://registry.npmjs.org/@octokit/graphql/-/graphql-7.1.1.tgz", @@ -2139,112 +1928,11 @@ "@octokit/openapi-types": "^24.2.0" } }, - "node_modules/@octokit/oauth-app": { - "version": "8.0.3", - "resolved": "https://registry.npmjs.org/@octokit/oauth-app/-/oauth-app-8.0.3.tgz", - "integrity": "sha512-jnAjvTsPepyUaMu9e69hYBuozEPgYqP4Z3UnpmvoIzHDpf8EXDGvTY1l1jK0RsZ194oRd+k6Hm13oRU8EoDFwg==", - "license": "MIT", - "dependencies": { - "@octokit/auth-oauth-app": "^9.0.2", - "@octokit/auth-oauth-user": "^6.0.1", - "@octokit/auth-unauthenticated": "^7.0.2", - "@octokit/core": "^7.0.5", - "@octokit/oauth-authorization-url": "^8.0.0", - "@octokit/oauth-methods": "^6.0.1", - "@types/aws-lambda": "^8.10.83", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/oauth-app/node_modules/@octokit/auth-token": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/@octokit/auth-token/-/auth-token-6.0.0.tgz", - "integrity": "sha512-P4YJBPdPSpWTQ1NU4XYdvHvXJJDxM6YwpS0FZHRgP7YFkdVxsWcpWGy/NVqlAA7PcPCnMacXlRm1y2PFZRWL/w==", - "license": "MIT", - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/oauth-app/node_modules/@octokit/core": { - "version": "7.0.6", - "resolved": "https://registry.npmjs.org/@octokit/core/-/core-7.0.6.tgz", - "integrity": "sha512-DhGl4xMVFGVIyMwswXeyzdL4uXD5OGILGX5N8Y+f6W7LhC1Ze2poSNrkF/fedpVDHEEZ+PHFW0vL14I+mm8K3Q==", - "license": "MIT", - "dependencies": { - "@octokit/auth-token": "^6.0.0", - "@octokit/graphql": "^9.0.3", - "@octokit/request": "^10.0.6", - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0", - "before-after-hook": "^4.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/oauth-app/node_modules/@octokit/graphql": { - "version": "9.0.3", - "resolved": "https://registry.npmjs.org/@octokit/graphql/-/graphql-9.0.3.tgz", - "integrity": "sha512-grAEuupr/C1rALFnXTv6ZQhFuL1D8G5y8CN04RgrO4FIPMrtm+mcZzFG7dcBm+nq+1ppNixu+Jd78aeJOYxlGA==", - "license": "MIT", - "dependencies": { - "@octokit/request": "^10.0.6", - "@octokit/types": "^16.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/oauth-app/node_modules/before-after-hook": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/before-after-hook/-/before-after-hook-4.0.0.tgz", - "integrity": "sha512-q6tR3RPqIB1pMiTRMFcZwuG5T8vwp+vUvEG0vuI6B+Rikh5BfPp2fQ82c925FOs+b0lcFQ8CFrL+KbilfZFhOQ==", - "license": "Apache-2.0" - }, - "node_modules/@octokit/oauth-app/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, - "node_modules/@octokit/oauth-authorization-url": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/@octokit/oauth-authorization-url/-/oauth-authorization-url-8.0.0.tgz", - "integrity": "sha512-7QoLPRh/ssEA/HuHBHdVdSgF8xNLz/Bc5m9fZkArJE5bb6NmVkDm3anKxXPmN1zh6b5WKZPRr3697xKT/yM3qQ==", - "license": "MIT", - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/oauth-methods": { - "version": "6.0.2", - "resolved": "https://registry.npmjs.org/@octokit/oauth-methods/-/oauth-methods-6.0.2.tgz", - "integrity": "sha512-HiNOO3MqLxlt5Da5bZbLV8Zarnphi4y9XehrbaFMkcoJ+FL7sMxH/UlUsCVxpddVu4qvNDrBdaTVE2o4ITK8ng==", - "license": "MIT", - "dependencies": { - "@octokit/oauth-authorization-url": "^8.0.0", - "@octokit/request": "^10.0.6", - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0" - }, - "engines": { - "node": ">= 20" - } - }, "node_modules/@octokit/openapi-types": { "version": "27.0.0", "resolved": "https://registry.npmjs.org/@octokit/openapi-types/-/openapi-types-27.0.0.tgz", "integrity": "sha512-whrdktVs1h6gtR+09+QsNk2+FO+49j6ga1c55YZudfEG+oKJVvJLQi3zkOm5JjiUXAagWK2tI2kTGKJ2Ys7MGA==", - "license": "MIT" - }, - "node_modules/@octokit/openapi-webhooks-types": { - "version": "12.0.3", - "resolved": "https://registry.npmjs.org/@octokit/openapi-webhooks-types/-/openapi-webhooks-types-12.0.3.tgz", - "integrity": "sha512-90MF5LVHjBedwoHyJsgmaFhEN1uzXyBDRLEBe7jlTYx/fEhPAk3P3DAJsfZwC54m8hAIryosJOL+UuZHB3K3yA==", + "dev": true, "license": "MIT" }, "node_modules/@octokit/plugin-request-log": { @@ -2298,72 +1986,16 @@ "@octokit/openapi-types": "^24.2.0" } }, - "node_modules/@octokit/request": { - "version": "10.0.6", - "resolved": "https://registry.npmjs.org/@octokit/request/-/request-10.0.6.tgz", - "integrity": "sha512-FO+UgZCUu+pPnZAR+iKdUt64kPE7QW7ciqpldaMXaNzixz5Jld8dJ31LAUewk0cfSRkNSRKyqG438ba9c/qDlQ==", - "license": "MIT", - "dependencies": { - "@octokit/endpoint": "^11.0.2", - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0", - "fast-content-type-parse": "^3.0.0", - "universal-user-agent": "^7.0.2" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/request-error": { - "version": "7.0.2", - "resolved": "https://registry.npmjs.org/@octokit/request-error/-/request-error-7.0.2.tgz", - "integrity": "sha512-U8piOROoQQUyExw5c6dTkU3GKxts5/ERRThIauNL7yaRoeXW0q/5bgHWT7JfWBw1UyrbK8ERId2wVkcB32n0uQ==", - "license": "MIT", - "dependencies": { - "@octokit/types": "^16.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/request/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, "node_modules/@octokit/types": { "version": "16.0.0", "resolved": "https://registry.npmjs.org/@octokit/types/-/types-16.0.0.tgz", "integrity": "sha512-sKq+9r1Mm4efXW1FCk7hFSeJo4QKreL/tTbR0rz/qx/r1Oa2VV83LTA/H/MuCOX7uCIJmQVRKBcbmWoySjAnSg==", + "dev": true, "license": "MIT", "dependencies": { "@octokit/openapi-types": "^27.0.0" } }, - "node_modules/@octokit/webhooks": { - "version": "14.1.3", - "resolved": "https://registry.npmjs.org/@octokit/webhooks/-/webhooks-14.1.3.tgz", - "integrity": "sha512-gcK4FNaROM9NjA0mvyfXl0KPusk7a1BeA8ITlYEZVQCXF5gcETTd4yhAU0Kjzd8mXwYHppzJBWgdBVpIR9wUcQ==", - "license": "MIT", - "dependencies": { - "@octokit/openapi-webhooks-types": "12.0.3", - "@octokit/request-error": "^7.0.0", - "@octokit/webhooks-methods": "^6.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/webhooks-methods": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/@octokit/webhooks-methods/-/webhooks-methods-6.0.0.tgz", - "integrity": "sha512-MFlzzoDJVw/GcbfzVC1RLR36QqkTLUf79vLVO3D+xn7r0QgxnFoLZgtrzxiQErAjFUOdH6fas2KeQJ1yr/qaXQ==", - "license": "MIT", - "engines": { - "node": ">= 20" - } - }, "node_modules/@open-draft/deferred-promise": { "version": "2.2.0", "resolved": "https://registry.npmjs.org/@open-draft/deferred-promise/-/deferred-promise-2.2.0.tgz", @@ -2612,12 +2244,6 @@ "@types/readdir-glob": "*" } }, - "node_modules/@types/aws-lambda": { - "version": "8.10.157", - "resolved": "https://registry.npmjs.org/@types/aws-lambda/-/aws-lambda-8.10.157.tgz", - "integrity": "sha512-ofjcRCO1N7tMZDSO11u5bFHPDfUFD3Q9YK9g4S4w8UDKuG3CNlw2lNK1sd3Itdo7JORygZmG4h9ZykS8dlXvMA==", - "license": "MIT" - }, "node_modules/@types/color-name": { "version": "1.1.1", "dev": true, @@ -5800,22 +5426,6 @@ "url": "https://github.com/sindresorhus/execa?sponsor=1" } }, - "node_modules/fast-content-type-parse": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/fast-content-type-parse/-/fast-content-type-parse-3.0.0.tgz", - "integrity": "sha512-ZvLdcY8P+N8mGQJahJV5G4U88CSvT1rP8ApL6uETe88MBXrBHAkZlSEySdUlyztF7ccb+Znos3TFqaepHxdhBg==", - "funding": [ - { - "type": "github", - "url": "https://github.com/sponsors/fastify" - }, - { - "type": "opencollective", - "url": "https://opencollective.com/fastify" - } - ], - "license": "MIT" - }, "node_modules/fast-deep-equal": { "version": "3.1.3", "license": "MIT" @@ -6171,7 +5781,6 @@ "version": "11.1.0", "resolved": "https://registry.npmjs.org/glob/-/glob-11.1.0.tgz", "integrity": "sha512-vuNwKSaKiqm7g0THUBu2x7ckSs3XJLXE+2ssL7/MfTGPLLcrJQ/4Uq1CjPTtO5cCIiRxqvN6Twy1qOwhL0Xjcw==", - "dev": true, "license": "BlueOak-1.0.0", "dependencies": { "foreground-child": "^3.3.1", @@ -6206,7 +5815,6 @@ "version": "10.1.1", "resolved": "https://registry.npmjs.org/minimatch/-/minimatch-10.1.1.tgz", "integrity": "sha512-enIvLvRAFZYXJzkCYG5RKmPfrFArdLv+R+lbQ53BmIMLIry74bjKzX6iHAm8WYamJkhSSEabrWN5D97XnKObjQ==", - "dev": true, "license": "BlueOak-1.0.0", "dependencies": { "@isaacs/brace-expansion": "^5.0.0" @@ -6875,7 +6483,6 @@ "version": "4.1.1", "resolved": "https://registry.npmjs.org/jackspeak/-/jackspeak-4.1.1.tgz", "integrity": "sha512-zptv57P3GpL+O0I7VdMJNBZCu+BPHVQUk55Ft8/QCJjTVxrnJHuVuX/0Bl2A6/+2oyR/ZMEuFKwmzqqZ/U5nPQ==", - "dev": true, "license": "BlueOak-1.0.0", "dependencies": { "@isaacs/cliui": "^8.0.2" @@ -6909,10 +6516,11 @@ } }, "node_modules/jschardet": { - "version": "3.1.3", - "resolved": "https://registry.npmjs.org/jschardet/-/jschardet-3.1.3.tgz", - "integrity": "sha512-Q1PKVMK/uu+yjdlobgWIYkUOCR1SqUmW9m/eUJNNj4zI2N12i25v8fYpVf+zCakQeaTdBdhnZTFbVIAVZIVVOg==", + "version": "3.1.4", + "resolved": "https://registry.npmjs.org/jschardet/-/jschardet-3.1.4.tgz", + "integrity": "sha512-/kmVISmrwVwtyYU40iQUOp3SUPk2dhNCMsZBQX0R1/jZ8maaXJ/oZIzUOiyOqcgtLnETFKYChbJ5iDC/eWmFHg==", "dev": true, + "license": "LGPL-2.1+", "engines": { "node": ">=0.1.90" } @@ -7118,7 +6726,6 @@ "version": "11.1.0", "resolved": "https://registry.npmjs.org/lru-cache/-/lru-cache-11.1.0.tgz", "integrity": "sha512-QIXZUBJUx+2zHUdQujWejBkcD9+cs94tLn0+YL8UrCh+D5sCXZ4c7LaEH48pNwRY3MLDgqUFyhlCyjJPf1WP0A==", - "dev": true, "license": "ISC", "engines": { "node": "20 || >=22" @@ -7509,153 +7116,6 @@ "url": "https://github.com/sponsors/ljharb" } }, - "node_modules/octokit": { - "version": "5.0.5", - "resolved": "https://registry.npmjs.org/octokit/-/octokit-5.0.5.tgz", - "integrity": "sha512-4+/OFSqOjoyULo7eN7EA97DE0Xydj/PW5aIckxqQIoFjFwqXKuFCvXUJObyJfBF9Khu4RL/jlDRI9FPaMGfPnw==", - "license": "MIT", - "dependencies": { - "@octokit/app": "^16.1.2", - "@octokit/core": "^7.0.6", - "@octokit/oauth-app": "^8.0.3", - "@octokit/plugin-paginate-graphql": "^6.0.0", - "@octokit/plugin-paginate-rest": "^14.0.0", - "@octokit/plugin-rest-endpoint-methods": "^17.0.0", - "@octokit/plugin-retry": "^8.0.3", - "@octokit/plugin-throttling": "^11.0.3", - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0", - "@octokit/webhooks": "^14.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/octokit/node_modules/@octokit/auth-token": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/@octokit/auth-token/-/auth-token-6.0.0.tgz", - "integrity": "sha512-P4YJBPdPSpWTQ1NU4XYdvHvXJJDxM6YwpS0FZHRgP7YFkdVxsWcpWGy/NVqlAA7PcPCnMacXlRm1y2PFZRWL/w==", - "engines": { - "node": ">= 20" - } - }, - "node_modules/octokit/node_modules/@octokit/core": { - "version": "7.0.6", - "resolved": "https://registry.npmjs.org/@octokit/core/-/core-7.0.6.tgz", - "integrity": "sha512-DhGl4xMVFGVIyMwswXeyzdL4uXD5OGILGX5N8Y+f6W7LhC1Ze2poSNrkF/fedpVDHEEZ+PHFW0vL14I+mm8K3Q==", - "license": "MIT", - "dependencies": { - "@octokit/auth-token": "^6.0.0", - "@octokit/graphql": "^9.0.3", - "@octokit/request": "^10.0.6", - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0", - "before-after-hook": "^4.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/octokit/node_modules/@octokit/graphql": { - "version": "9.0.3", - "resolved": "https://registry.npmjs.org/@octokit/graphql/-/graphql-9.0.3.tgz", - "integrity": "sha512-grAEuupr/C1rALFnXTv6ZQhFuL1D8G5y8CN04RgrO4FIPMrtm+mcZzFG7dcBm+nq+1ppNixu+Jd78aeJOYxlGA==", - "license": "MIT", - "dependencies": { - "@octokit/request": "^10.0.6", - "@octokit/types": "^16.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/octokit/node_modules/@octokit/plugin-paginate-graphql": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/@octokit/plugin-paginate-graphql/-/plugin-paginate-graphql-6.0.0.tgz", - "integrity": "sha512-crfpnIoFiBtRkvPqOyLOsw12XsveYuY2ieP6uYDosoUegBJpSVxGwut9sxUgFFcll3VTOTqpUf8yGd8x1OmAkQ==", - "engines": { - "node": ">= 20" - }, - "peerDependencies": { - "@octokit/core": ">=6" - } - }, - "node_modules/octokit/node_modules/@octokit/plugin-paginate-rest": { - "version": "14.0.0", - "resolved": "https://registry.npmjs.org/@octokit/plugin-paginate-rest/-/plugin-paginate-rest-14.0.0.tgz", - "integrity": "sha512-fNVRE7ufJiAA3XUrha2omTA39M6IXIc6GIZLvlbsm8QOQCYvpq/LkMNGyFlB1d8hTDzsAXa3OKtybdMAYsV/fw==", - "license": "MIT", - "dependencies": { - "@octokit/types": "^16.0.0" - }, - "engines": { - "node": ">= 20" - }, - "peerDependencies": { - "@octokit/core": ">=6" - } - }, - "node_modules/octokit/node_modules/@octokit/plugin-rest-endpoint-methods": { - "version": "17.0.0", - "resolved": "https://registry.npmjs.org/@octokit/plugin-rest-endpoint-methods/-/plugin-rest-endpoint-methods-17.0.0.tgz", - "integrity": "sha512-B5yCyIlOJFPqUUeiD0cnBJwWJO8lkJs5d8+ze9QDP6SvfiXSz1BF+91+0MeI1d2yxgOhU/O+CvtiZ9jSkHhFAw==", - "license": "MIT", - "dependencies": { - "@octokit/types": "^16.0.0" - }, - "engines": { - "node": ">= 20" - }, - "peerDependencies": { - "@octokit/core": ">=6" - } - }, - "node_modules/octokit/node_modules/@octokit/plugin-retry": { - "version": "8.0.3", - "resolved": "https://registry.npmjs.org/@octokit/plugin-retry/-/plugin-retry-8.0.3.tgz", - "integrity": "sha512-vKGx1i3MC0za53IzYBSBXcrhmd+daQDzuZfYDd52X5S0M2otf3kVZTVP8bLA3EkU0lTvd1WEC2OlNNa4G+dohA==", - "license": "MIT", - "dependencies": { - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0", - "bottleneck": "^2.15.3" - }, - "engines": { - "node": ">= 20" - }, - "peerDependencies": { - "@octokit/core": ">=7" - } - }, - "node_modules/octokit/node_modules/@octokit/plugin-throttling": { - "version": "11.0.3", - "resolved": "https://registry.npmjs.org/@octokit/plugin-throttling/-/plugin-throttling-11.0.3.tgz", - "integrity": "sha512-34eE0RkFCKycLl2D2kq7W+LovheM/ex3AwZCYN8udpi6bxsyjZidb2McXs69hZhLmJlDqTSP8cH+jSRpiaijBg==", - "license": "MIT", - "dependencies": { - "@octokit/types": "^16.0.0", - "bottleneck": "^2.15.3" - }, - "engines": { - "node": ">= 20" - }, - "peerDependencies": { - "@octokit/core": "^7.0.0" - } - }, - "node_modules/octokit/node_modules/before-after-hook": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/before-after-hook/-/before-after-hook-4.0.0.tgz", - "integrity": "sha512-q6tR3RPqIB1pMiTRMFcZwuG5T8vwp+vUvEG0vuI6B+Rikh5BfPp2fQ82c925FOs+b0lcFQ8CFrL+KbilfZFhOQ==" - }, - "node_modules/octokit/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, "node_modules/once": { "version": "1.4.0", "license": "ISC", @@ -7776,7 +7236,6 @@ "version": "2.0.0", "resolved": "https://registry.npmjs.org/path-scurry/-/path-scurry-2.0.0.tgz", "integrity": "sha512-ypGJsmGtdXUOeM5u93TyeIEfEhM6s+ljAhrk5vAvSx8uyY/02OvrZnA0YNGUrPXfpJMgI1ODd3nwz8Npx4O4cg==", - "dev": true, "license": "BlueOak-1.0.0", "dependencies": { "lru-cache": "^11.0.0", @@ -8900,15 +8359,6 @@ "url": "https://github.com/sponsors/sindresorhus" } }, - "node_modules/toad-cache": { - "version": "3.7.0", - "resolved": "https://registry.npmjs.org/toad-cache/-/toad-cache-3.7.0.tgz", - "integrity": "sha512-/m8M+2BJUpoJdgAHoG+baCwBT+tf2VraSfkBgl0Y00qIWt41DJ8R5B8nsEw0I58YwF5IZH6z24/2TobDKnqSWw==", - "license": "MIT", - "engines": { - "node": ">=12" - } - }, "node_modules/tr46": { "version": "0.0.3", "license": "MIT" @@ -9231,12 +8681,6 @@ "url": "https://github.com/sponsors/sindresorhus" } }, - "node_modules/universal-github-app-jwt": { - "version": "2.2.2", - "resolved": "https://registry.npmjs.org/universal-github-app-jwt/-/universal-github-app-jwt-2.2.2.tgz", - "integrity": "sha512-dcmbeSrOdTnsjGjUfAlqNDJrhxXizjAz94ija9Qw8YkZ1uu0d+GoZzyH+Jb9tIIqvGsadUfwg+22k5aDqqwzbw==", - "license": "MIT" - }, "node_modules/universal-user-agent": { "version": "6.0.0", "license": "ISC" @@ -9294,7 +8738,8 @@ "version": "3.0.0", "resolved": "https://registry.npmjs.org/utf8/-/utf8-3.0.0.tgz", "integrity": "sha512-E8VjFIQ/TyQgp+TZfS6l8yp/xWppSAHzidGiRrqe4bK4XP9pTRyKFgGJpO3SN7zdX4DeomTrwaseCHovfpFcqQ==", - "dev": true + "dev": true, + "license": "MIT" }, "node_modules/util-deprecate": { "version": "1.0.2", diff --git a/package.json b/package.json index 6eedf7f47..a4064dea9 100644 --- a/package.json +++ b/package.json @@ -35,7 +35,6 @@ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", "archiver": "^7.0.1", "fast-deep-equal": "^3.1.3", @@ -45,7 +44,6 @@ "jsonschema": "1.4.1", "long": "^5.3.2", "node-forge": "^1.3.1", - "octokit": "^5.0.5", "semver": "^7.7.3", "uuid": "^13.0.0" }, @@ -59,7 +57,7 @@ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", "@types/sinon": "^17.0.4", @@ -67,13 +65,13 @@ "@typescript-eslint/parser": "^8.41.0", "ava": "^6.4.1", "esbuild": "^0.27.0", + "eslint": "^8.57.1", "eslint-import-resolver-typescript": "^3.8.7", "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", "eslint-plugin-jsdoc": "^61.1.12", "eslint-plugin-no-async-foreach": "^0.1.1", - "eslint": "^8.57.1", "glob": "^11.1.0", "nock": "^14.0.10", "sinon": "^21.0.0", From cac5926de5e2c542638262c4e24527972023cdec Mon Sep 17 00:00:00 2001 From: Henry Mercer Date: Tue, 18 Nov 2025 18:10:09 +0000 Subject: [PATCH 36/51] Delete unused exports --- lib/analyze-action-post.js | 25 +++++----- lib/analyze-action.js | 25 +++++----- lib/autobuild-action.js | 25 +++++----- lib/init-action-post.js | 25 +++++----- lib/init-action.js | 25 +++++----- lib/resolve-environment-action.js | 25 +++++----- lib/setup-codeql-action.js | 25 +++++----- lib/start-proxy-action-post.js | 25 +++++----- lib/start-proxy-action.js | 25 +++++----- lib/upload-lib.js | 27 +++++----- lib/upload-sarif-action-post.js | 25 +++++----- lib/upload-sarif-action.js | 25 +++++----- src/actions-util.ts | 2 +- src/analyses.ts | 2 +- src/api-client.ts | 5 -- src/cli-errors.ts | 5 +- src/codeql.ts | 4 +- src/config-utils.ts | 8 +-- src/dependency-caching.ts | 2 +- src/environment.ts | 14 ------ src/git-utils.ts | 61 ----------------------- src/overlay-database-utils.ts | 2 +- src/setup-codeql.ts | 23 +-------- src/start-proxy.ts | 4 +- src/status-report.ts | 2 +- src/tools-download.ts | 2 +- src/tracer-config.ts | 2 +- src/upload-lib.ts | 2 +- src/util.ts | 83 ------------------------------- 29 files changed, 163 insertions(+), 362 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index 6da36a092..4f4f64b0a 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -33934,7 +33934,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core14 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob2 = __importStar4(require_glob()); var io6 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -34018,7 +34018,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core14.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -116554,7 +116554,6 @@ var io2 = __toESM(require_io2()); // src/util.ts var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 22666ed2a..ff0b4db29 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -33934,7 +33934,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core15 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob2 = __importStar4(require_glob()); var io7 = __importStar4(require_io3()); var crypto2 = __importStar4(require("crypto")); @@ -34018,7 +34018,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core15.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -84336,7 +84336,6 @@ var fsPromises = __toESM(require("fs/promises")); var os = __toESM(require("os")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index 19055f637..725f9c8e1 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -33934,7 +33934,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core14 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob = __importStar4(require_glob()); var io5 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -34018,7 +34018,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core14.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -80332,7 +80332,6 @@ var io2 = __toESM(require_io2()); var fsPromises = __toESM(require("fs/promises")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js diff --git a/lib/init-action-post.js b/lib/init-action-post.js index 5857024ff..298746d93 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -33934,7 +33934,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core18 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob2 = __importStar4(require_glob()); var io7 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -34018,7 +34018,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core18.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -119452,7 +119452,6 @@ var fs = __toESM(require("fs")); var fsPromises = __toESM(require("fs/promises")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js diff --git a/lib/init-action.js b/lib/init-action.js index 267a67e31..c3199ca3f 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -34085,7 +34085,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core14 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob2 = __importStar4(require_glob()); var io7 = __importStar4(require_io3()); var crypto2 = __importStar4(require("crypto")); @@ -34169,7 +34169,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core14.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -81641,7 +81641,6 @@ var fsPromises = __toESM(require("fs/promises")); var os = __toESM(require("os")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index d5769de3a..3a4dd9efd 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -33934,7 +33934,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core13 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob = __importStar4(require_glob()); var io5 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -34018,7 +34018,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core13.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -80332,7 +80332,6 @@ var io2 = __toESM(require_io2()); var fsPromises = __toESM(require("fs/promises")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index b9577ce2a..eebe3d104 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -32637,7 +32637,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core13 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob = __importStar4(require_glob()); var io6 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -32721,7 +32721,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core13.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -80388,7 +80388,6 @@ var fs = __toESM(require("fs")); var fsPromises = __toESM(require("fs/promises")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index 894cec283..59854ff22 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -33934,7 +33934,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core14 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob2 = __importStar4(require_glob()); var io6 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -34018,7 +34018,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core14.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -116551,7 +116551,6 @@ var io2 = __toESM(require_io2()); // src/util.ts var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index 378fe7c8a..e23d99449 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -53592,7 +53592,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core12 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob = __importStar4(require_glob()); var io4 = __importStar4(require_io4()); var crypto = __importStar4(require("crypto")); @@ -53676,7 +53676,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core12.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -96767,7 +96767,6 @@ var io2 = __toESM(require_io3()); // src/util.ts var fsPromises = __toESM(require("fs/promises")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io3()); // node_modules/get-folder-size/index.js diff --git a/lib/upload-lib.js b/lib/upload-lib.js index a0881cc10..6f5d62a8c 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -33934,7 +33934,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core12 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob = __importStar4(require_glob()); var io6 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -34018,7 +34018,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core12.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -83221,7 +83221,6 @@ __export(upload_lib_exports, { buildPayload: () => buildPayload, findSarifFilesInDir: () => findSarifFilesInDir, getGroupedSarifFilePaths: () => getGroupedSarifFilePaths, - getSarifFilePaths: () => getSarifFilePaths, populateRunAutomationDetails: () => populateRunAutomationDetails, postProcessSarifFiles: () => postProcessSarifFiles, readSarifFile: () => readSarifFile, @@ -83257,7 +83256,6 @@ var io2 = __toESM(require_io2()); var fs = __toESM(require("fs")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js @@ -90652,7 +90650,6 @@ function filterAlertsByDiffRange(logger, sarif) { buildPayload, findSarifFilesInDir, getGroupedSarifFilePaths, - getSarifFilePaths, populateRunAutomationDetails, postProcessSarifFiles, readSarifFile, diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index 55a3abf8c..8f64f60ef 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -108407,7 +108407,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core14 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob2 = __importStar4(require_glob2()); var io6 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -108491,7 +108491,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core14.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -116551,7 +116551,6 @@ var io2 = __toESM(require_io2()); // src/util.ts var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 25f9547cc..8a115e6b4 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -32637,7 +32637,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core14 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob = __importStar4(require_glob()); var io6 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -32721,7 +32721,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core14.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -83230,7 +83230,6 @@ var fs = __toESM(require("fs")); var fsPromises = __toESM(require("fs/promises")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js diff --git a/src/actions-util.ts b/src/actions-util.ts index a2d691b42..736d35d0f 100644 --- a/src/actions-util.ts +++ b/src/actions-util.ts @@ -80,7 +80,7 @@ export function isRunningLocalAction(): boolean { * * This can be used to get the Action's name or tell if we're running a local Action. */ -export function getRelativeScriptPath(): string { +function getRelativeScriptPath(): string { const runnerTemp = getRequiredEnvParam("RUNNER_TEMP"); const actionsDirectory = path.join(path.dirname(runnerTemp), "_actions"); return path.relative(actionsDirectory, __filename); diff --git a/src/analyses.ts b/src/analyses.ts index a8d172e3b..4f91ab07c 100644 --- a/src/analyses.ts +++ b/src/analyses.ts @@ -98,7 +98,7 @@ export async function getAnalysisKinds( export const codeQualityQueries: string[] = ["code-quality"]; // Enumerates API endpoints that accept SARIF files. -export enum SARIF_UPLOAD_ENDPOINT { +enum SARIF_UPLOAD_ENDPOINT { CODE_SCANNING = "PUT /repos/:owner/:repo/code-scanning/analysis", CODE_QUALITY = "PUT /repos/:owner/:repo/code-quality/analysis", } diff --git a/src/api-client.ts b/src/api-client.ts index e14048337..600da1ed6 100644 --- a/src/api-client.ts +++ b/src/api-client.ts @@ -18,11 +18,6 @@ import { const GITHUB_ENTERPRISE_VERSION_HEADER = "x-github-enterprise-version"; -export enum DisallowedAPIVersionReason { - ACTION_TOO_OLD, - ACTION_TOO_NEW, -} - export type GitHubApiCombinedDetails = GitHubApiDetails & GitHubApiExternalRepoDetails; diff --git a/src/cli-errors.ts b/src/cli-errors.ts index 0a009f983..5aba268ca 100644 --- a/src/cli-errors.ts +++ b/src/cli-errors.ts @@ -159,10 +159,7 @@ type CliErrorConfiguration = { * All of our caught CLI error messages that we handle specially: ie. if we * would like to categorize an error as a configuration error or not. */ -export const cliErrorsConfig: Record< - CliConfigErrorCategory, - CliErrorConfiguration -> = { +const cliErrorsConfig: Record = { [CliConfigErrorCategory.AutobuildError]: { cliErrorMessageCandidates: [ new RegExp("We were unable to automatically build your code"), diff --git a/src/codeql.ts b/src/codeql.ts index bc1d00401..17db0ef2c 100644 --- a/src/codeql.ts +++ b/src/codeql.ts @@ -513,7 +513,7 @@ export async function getCodeQLForTesting( * version requirement. Must be set to true outside tests. * @returns A new CodeQL object */ -export async function getCodeQLForCmd( +async function getCodeQLForCmd( cmd: string, checkVersion: boolean, ): Promise { @@ -1222,7 +1222,7 @@ export async function getTrapCachingExtractorConfigArgsForLang( * * This will not exist if the configuration is being parsed in the Action. */ -export function getGeneratedCodeScanningConfigPath(config: Config): string { +function getGeneratedCodeScanningConfigPath(config: Config): string { return path.resolve(config.tempDir, "user-config.yaml"); } diff --git a/src/config-utils.ts b/src/config-utils.ts index 03608fb1a..016bc4868 100644 --- a/src/config-utils.ts +++ b/src/config-utils.ts @@ -179,7 +179,7 @@ export interface Config { repositoryProperties: RepositoryProperties; } -export async function getSupportedLanguageMap( +async function getSupportedLanguageMap( codeql: CodeQL, logger: Logger, ): Promise> { @@ -242,7 +242,7 @@ export function hasActionsWorkflows(sourceRoot: string): boolean { /** * Gets the set of languages in the current repository. */ -export async function getRawLanguagesInRepo( +async function getRawLanguagesInRepo( repository: RepositoryNwo, sourceRoot: string, logger: Logger, @@ -351,7 +351,7 @@ export function getRawLanguagesNoAutodetect( * @returns A tuple containing a list of languages in this repository that might be * analyzable and whether or not this list was determined automatically. */ -export async function getRawLanguages( +async function getRawLanguages( languagesInput: string | undefined, repository: RepositoryNwo, sourceRoot: string, @@ -1230,7 +1230,7 @@ export function isCodeQualityEnabled(config: Config): boolean { * @returns Returns `AnalysisKind.CodeScanning` if `AnalysisKind.CodeScanning` is enabled; * otherwise `AnalysisKind.CodeQuality`. */ -export function getPrimaryAnalysisKind(config: Config): AnalysisKind { +function getPrimaryAnalysisKind(config: Config): AnalysisKind { return isCodeScanningEnabled(config) ? AnalysisKind.CodeScanning : AnalysisKind.CodeQuality; diff --git a/src/dependency-caching.ts b/src/dependency-caching.ts index 9f6aa2afc..350d8e468 100644 --- a/src/dependency-caching.ts +++ b/src/dependency-caching.ts @@ -55,7 +55,7 @@ export function getJavaTempDependencyDir(): string { * @returns The paths of directories on the runner that should be included in a dependency cache * for a Java analysis. */ -export function getJavaDependencyDirs(): string[] { +function getJavaDependencyDirs(): string[] { return [ // Maven join(os.homedir(), ".m2", "repository"), diff --git a/src/environment.ts b/src/environment.ts index 16a016aaa..1d33c68a6 100644 --- a/src/environment.ts +++ b/src/environment.ts @@ -20,12 +20,6 @@ export enum EnvVar { /** Whether the CodeQL Action has invoked the Go autobuilder. */ DID_AUTOBUILD_GOLANG = "CODEQL_ACTION_DID_AUTOBUILD_GOLANG", - /** - * Whether to disable the SARIF post-processing in the Action that removes duplicate locations from - * notifications in the `run[].invocations[].toolExecutionNotifications` SARIF property. - */ - DISABLE_DUPLICATE_LOCATION_FIX = "CODEQL_ACTION_DISABLE_DUPLICATE_LOCATION_FIX", - /** * Whether the CodeQL Action is using its own deprecated and non-standard way of scanning for * multiple languages. @@ -56,20 +50,12 @@ export enum EnvVar { /** Whether the error for a deprecated version of the CodeQL Action was logged. */ LOG_VERSION_DEPRECATION = "CODEQL_ACTION_DID_LOG_VERSION_DEPRECATION", - /** - * For macOS. Result of `csrutil status` to determine whether System Integrity - * Protection is enabled. - */ - IS_SIP_ENABLED = "CODEQL_ACTION_IS_SIP_ENABLED", - /** UUID representing the current job run. */ JOB_RUN_UUID = "JOB_RUN_UUID", /** Status for the entire job, submitted to the status report in `init-post` */ JOB_STATUS = "CODEQL_ACTION_JOB_STATUS", - ODASA_TRACER_CONFIGURATION = "ODASA_TRACER_CONFIGURATION", - /** The value of the `output` input for the analyze action. */ SARIF_RESULTS_OUTPUT_DIR = "CODEQL_ACTION_SARIF_RESULTS_OUTPUT_DIR", diff --git a/src/git-utils.ts b/src/git-utils.ts index 837124027..0d2a7df7a 100644 --- a/src/git-utils.ts +++ b/src/git-utils.ts @@ -122,67 +122,6 @@ export const determineBaseBranchHeadCommitOid = async function ( } }; -/** - * Deepen the git history of HEAD by one level. Errors are logged. - * - * This function uses the `checkout_path` to determine the repository path and - * works only when called from `analyze` or `upload-sarif`. - */ -export const deepenGitHistory = async function () { - try { - await runGitCommand( - getOptionalInput("checkout_path"), - [ - "fetch", - "origin", - "HEAD", - "--no-tags", - "--no-recurse-submodules", - "--deepen=1", - ], - "Cannot deepen the shallow repository.", - ); - } catch { - // Errors are already logged by runGitCommand() - } -}; - -/** - * Fetch the given remote branch. Errors are logged. - * - * This function uses the `checkout_path` to determine the repository path and - * works only when called from `analyze` or `upload-sarif`. - */ -export const gitFetch = async function (branch: string, extraFlags: string[]) { - try { - await runGitCommand( - getOptionalInput("checkout_path"), - ["fetch", "--no-tags", ...extraFlags, "origin", `${branch}:${branch}`], - `Cannot fetch ${branch}.`, - ); - } catch { - // Errors are already logged by runGitCommand() - } -}; - -/** - * Repack the git repository, using with the given flags. Errors are logged. - * - * This function uses the `checkout_path` to determine the repository path and - * works only when called from `analyze` or `upload-sarif`. - */ -export const gitRepack = async function (flags: string[]) { - try { - await runGitCommand( - getOptionalInput("checkout_path"), - ["repack", ...flags], - "Cannot repack the repository.", - ); - } catch { - // Errors are already logged by runGitCommand() - } -}; - /** * Decode, if necessary, a file path produced by Git. See * https://git-scm.com/docs/git-config#Documentation/git-config.txt-corequotePath diff --git a/src/overlay-database-utils.ts b/src/overlay-database-utils.ts index ebb020ba8..71990ccc3 100644 --- a/src/overlay-database-utils.ts +++ b/src/overlay-database-utils.ts @@ -175,7 +175,7 @@ const MAX_CACHE_OPERATION_MS = 600_000; * @param warningPrefix Prefix for the check failure warning message * @returns True if the verification succeeded, false otherwise */ -export function checkOverlayBaseDatabase( +function checkOverlayBaseDatabase( config: Config, logger: Logger, warningPrefix: string, diff --git a/src/setup-codeql.ts b/src/setup-codeql.ts index 948893022..6a412d1df 100644 --- a/src/setup-codeql.ts +++ b/src/setup-codeql.ts @@ -34,7 +34,7 @@ export enum ToolsSource { Download = "DOWNLOAD", } -export const CODEQL_DEFAULT_ACTION_REPOSITORY = "github/codeql-action"; +const CODEQL_DEFAULT_ACTION_REPOSITORY = "github/codeql-action"; const CODEQL_NIGHTLIES_REPOSITORY_OWNER = "dsp-testing"; const CODEQL_NIGHTLIES_REPOSITORY_NAME = "codeql-cli-nightlies"; @@ -180,17 +180,6 @@ export function tryGetTagNameFromUrl( return match[1]; } -export function tryGetBundleVersionFromUrl( - url: string, - logger: Logger, -): string | undefined { - const tagName = tryGetTagNameFromUrl(url, logger); - if (tagName === undefined) { - return undefined; - } - return tryGetBundleVersionFromTagName(tagName, logger); -} - export function convertToSemVer(version: string, logger: Logger): string { if (!semver.valid(version)) { logger.debug( @@ -580,7 +569,7 @@ export async function getCodeQLSource( * Gets a fallback version number to use when looking for CodeQL in the toolcache if we didn't find * the `x.y.z` version. This is to support old versions of the toolcache. */ -export async function tryGetFallbackToolcacheVersion( +async function tryGetFallbackToolcacheVersion( cliVersion: string | undefined, tagName: string, logger: Logger, @@ -729,14 +718,6 @@ function getCanonicalToolcacheVersion( return cliVersion; } -export interface SetupCodeQLResult { - codeqlFolder: string; - toolsDownloadStatusReport?: ToolsDownloadStatusReport; - toolsSource: ToolsSource; - toolsVersion: string; - zstdAvailability: tar.ZstdAvailability; -} - /** * Obtains the CodeQL bundle, installs it in the toolcache if appropriate, and extracts it. * diff --git a/src/start-proxy.ts b/src/start-proxy.ts index 2888e1a58..2a082ed62 100644 --- a/src/start-proxy.ts +++ b/src/start-proxy.ts @@ -8,7 +8,7 @@ import { ConfigurationError, getErrorMessage, isDefined } from "./util"; export const UPDATEJOB_PROXY = "update-job-proxy"; export const UPDATEJOB_PROXY_VERSION = "v2.0.20250624110901"; -export const UPDATEJOB_PROXY_URL_PREFIX = +const UPDATEJOB_PROXY_URL_PREFIX = "https://github.com/github/codeql-action/releases/download/codeql-bundle-v2.22.0/"; export type Credential = { @@ -202,7 +202,7 @@ export function getFallbackUrl(proxyPackage: string): string { * * @returns The response from the GitHub API. */ -export async function getLinkedRelease() { +async function getLinkedRelease() { return getApiClient().rest.repos.getReleaseByTag({ owner: "github", repo: "codeql-action", diff --git a/src/status-report.ts b/src/status-report.ts index 1ad53ac32..c6e747489 100644 --- a/src/status-report.ts +++ b/src/status-report.ts @@ -54,7 +54,7 @@ export enum ActionName { * considered to be a third party analysis and is treated differently when calculating SLOs. To ensure * misconfigured workflows are not treated as third party, only the upload-sarif action can return false. */ -export function isFirstPartyAnalysis(actionName: ActionName): boolean { +function isFirstPartyAnalysis(actionName: ActionName): boolean { if (actionName !== ActionName.UploadSarif) { return true; } diff --git a/src/tools-download.ts b/src/tools-download.ts index 4cfba397e..5d8a4c5fb 100644 --- a/src/tools-download.ts +++ b/src/tools-download.ts @@ -17,7 +17,7 @@ import { cleanUpPath, getErrorMessage, getRequiredEnvParam } from "./util"; /** * High watermark to use when streaming the download and extraction of the CodeQL tools. */ -export const STREAMING_HIGH_WATERMARK_BYTES = 4 * 1024 * 1024; // 4 MiB +const STREAMING_HIGH_WATERMARK_BYTES = 4 * 1024 * 1024; // 4 MiB /** * The name of the tool cache directory for the CodeQL tools. diff --git a/src/tracer-config.ts b/src/tracer-config.ts index 9eea3eecc..d786d4651 100644 --- a/src/tracer-config.ts +++ b/src/tracer-config.ts @@ -76,7 +76,7 @@ export async function endTracingForCluster( } } -export async function getTracerConfigForCluster( +async function getTracerConfigForCluster( config: Config, ): Promise { const tracingEnvVariables = JSON.parse( diff --git a/src/upload-lib.ts b/src/upload-lib.ts index f032b8327..ac0274520 100644 --- a/src/upload-lib.ts +++ b/src/upload-lib.ts @@ -412,7 +412,7 @@ export function findSarifFilesInDir( return sarifFiles; } -export function getSarifFilePaths( +function getSarifFilePaths( sarifPath: string, isSarif: (name: string) => boolean, ) { diff --git a/src/util.ts b/src/util.ts index f23c3be7d..fe8604b46 100644 --- a/src/util.ts +++ b/src/util.ts @@ -4,7 +4,6 @@ import * as os from "os"; import * as path from "path"; import * as core from "@actions/core"; -import * as exec from "@actions/exec/lib/exec"; import * as io from "@actions/io"; import getFolderSize from "get-folder-size"; import * as yaml from "js-yaml"; @@ -1026,34 +1025,6 @@ export function fixInvalidNotifications( return newSarif; } -/** - * Removes duplicates from the sarif file. - * - * When `CODEQL_ACTION_DISABLE_DUPLICATE_LOCATION_FIX` is set to true, this will - * simply rename the input file to the output file. Otherwise, it will parse the - * input file as JSON, remove duplicate locations from the SARIF notification - * objects, and write the result to the output file. - * - * For context, see documentation of: - * `CODEQL_ACTION_DISABLE_DUPLICATE_LOCATION_FIX`. */ -export function fixInvalidNotificationsInFile( - inputPath: string, - outputPath: string, - logger: Logger, -): void { - if (process.env[EnvVar.DISABLE_DUPLICATE_LOCATION_FIX] === "true") { - logger.info( - "SARIF notification object duplicate location fix disabled by the " + - `${EnvVar.DISABLE_DUPLICATE_LOCATION_FIX} environment variable.`, - ); - fs.renameSync(inputPath, outputPath); - } else { - let sarif = JSON.parse(fs.readFileSync(inputPath, "utf8")) as SarifFile; - sarif = fixInvalidNotifications(sarif, logger); - fs.writeFileSync(outputPath, JSON.stringify(sarif)); - } -} - export function wrapError(error: unknown): Error { return error instanceof Error ? error : new Error(String(error)); } @@ -1197,49 +1168,6 @@ export function cloneObject(obj: T): T { return JSON.parse(JSON.stringify(obj)) as T; } -// The first time this function is called, it runs `csrutil status` to determine -// whether System Integrity Protection is enabled; and saves the result in an -// environment variable. Afterwards, simply return the value of the environment -// variable. -export async function checkSipEnablement( - logger: Logger, -): Promise { - if ( - process.env[EnvVar.IS_SIP_ENABLED] !== undefined && - ["true", "false"].includes(process.env[EnvVar.IS_SIP_ENABLED]) - ) { - return process.env[EnvVar.IS_SIP_ENABLED] === "true"; - } - - try { - const sipStatusOutput = await exec.getExecOutput("csrutil status"); - if (sipStatusOutput.exitCode === 0) { - if ( - sipStatusOutput.stdout.includes( - "System Integrity Protection status: enabled.", - ) - ) { - core.exportVariable(EnvVar.IS_SIP_ENABLED, "true"); - return true; - } - if ( - sipStatusOutput.stdout.includes( - "System Integrity Protection status: disabled.", - ) - ) { - core.exportVariable(EnvVar.IS_SIP_ENABLED, "false"); - return false; - } - } - return undefined; - } catch (e) { - logger.warning( - `Failed to determine if System Integrity Protection was enabled: ${e}`, - ); - return undefined; - } -} - export async function cleanUpPath(file: string, name: string, logger: Logger) { logger.debug(`Cleaning up ${name}.`); try { @@ -1291,17 +1219,6 @@ export function isDefined(value: T | null | undefined): value is T { return value !== undefined && value !== null; } -/** Like `Object.keys`, but typed so that the elements of the resulting array have the - * same type as the keys of the input object. Note that this may not be sound if the input - * object has been cast to `T` from a subtype of `T` and contains additional keys that - * are not represented by `keyof T`. - */ -export function unsafeKeysInvariant>( - object: T, -): Array { - return Object.keys(object) as Array; -} - /** Like `Object.entries`, but typed so that the key elements of the result have the * same type as the keys of the input object. Note that this may not be sound if the input * object has been cast to `T` from a subtype of `T` and contains additional keys that From 5da2098551dbd40676c16f35ab142315ccbd1fba Mon Sep 17 00:00:00 2001 From: Henry Mercer Date: Tue, 18 Nov 2025 15:19:18 +0000 Subject: [PATCH 37/51] Add feature flag for uploading overlay DBs to API --- lib/analyze-action-post.js | 21 +++++++++++++-------- lib/analyze-action.js | 21 +++++++++++++-------- lib/autobuild-action.js | 21 +++++++++++++-------- lib/init-action-post.js | 21 +++++++++++++-------- lib/init-action.js | 21 +++++++++++++-------- lib/resolve-environment-action.js | 21 +++++++++++++-------- lib/setup-codeql-action.js | 21 +++++++++++++-------- lib/start-proxy-action-post.js | 21 +++++++++++++-------- lib/start-proxy-action.js | 21 +++++++++++++-------- lib/upload-lib.js | 21 +++++++++++++-------- lib/upload-sarif-action-post.js | 21 +++++++++++++-------- lib/upload-sarif-action.js | 21 +++++++++++++-------- package-lock.json | 7 ------- src/feature-flags.ts | 22 ++++++++++++++-------- 14 files changed, 170 insertions(+), 111 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index 7dde941ee..6b3eb444f 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -120074,6 +120074,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -120185,21 +120190,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 242a05aca..a2fe4e0b0 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -88695,6 +88695,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -88806,21 +88811,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index 3049d15a2..e4ba8a19f 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -84014,6 +84014,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -84125,21 +84130,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, diff --git a/lib/init-action-post.js b/lib/init-action-post.js index 5dd89dcf6..112bb5854 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -123455,6 +123455,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -123566,21 +123571,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, diff --git a/lib/init-action.js b/lib/init-action.js index 4d68e5c9a..e052ffd70 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -86109,6 +86109,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -86220,21 +86225,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index c3d54f680..9db9daed4 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -84005,6 +84005,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -84116,21 +84121,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index 973e9c431..fefd1fedb 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -83917,6 +83917,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -84028,21 +84033,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index 7e34a5e95..066ceed28 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -119480,6 +119480,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -119591,21 +119596,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index c0869dd96..65bbb9a5e 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -100033,6 +100033,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -100144,21 +100149,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, diff --git a/lib/upload-lib.js b/lib/upload-lib.js index ea6e2ca41..ded95017f 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -87071,6 +87071,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -87182,21 +87187,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index fce0c4f79..abe7dcbd7 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -119646,6 +119646,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -119757,21 +119762,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 667acaa15..5b83db2a6 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -86867,6 +86867,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -86978,21 +86983,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, diff --git a/package-lock.json b/package-lock.json index 80ef7cb61..e7fd2894d 100644 --- a/package-lock.json +++ b/package-lock.json @@ -1578,7 +1578,6 @@ "version": "4.0.1", "resolved": "https://registry.npmjs.org/@isaacs/balanced-match/-/balanced-match-4.0.1.tgz", "integrity": "sha512-yzMTt9lEb8Gv7zRioUilSglI0c0smZ9k5D65677DLWLtWJaXIS3CqcGyUFByYKlnUj6TkjLVs54fBl6+TiGQDQ==", - "dev": true, "license": "MIT", "engines": { "node": "20 || >=22" @@ -1588,7 +1587,6 @@ "version": "5.0.0", "resolved": "https://registry.npmjs.org/@isaacs/brace-expansion/-/brace-expansion-5.0.0.tgz", "integrity": "sha512-ZT55BDLV0yv0RBm2czMiZ+SqCGO7AvmOM3G/w2xhVPH+te0aKgFjmBvGlL1dH+ql2tgGO3MVrbb3jCKyvpgnxA==", - "dev": true, "license": "MIT", "dependencies": { "@isaacs/balanced-match": "^4.0.1" @@ -6171,7 +6169,6 @@ "version": "11.1.0", "resolved": "https://registry.npmjs.org/glob/-/glob-11.1.0.tgz", "integrity": "sha512-vuNwKSaKiqm7g0THUBu2x7ckSs3XJLXE+2ssL7/MfTGPLLcrJQ/4Uq1CjPTtO5cCIiRxqvN6Twy1qOwhL0Xjcw==", - "dev": true, "license": "BlueOak-1.0.0", "dependencies": { "foreground-child": "^3.3.1", @@ -6206,7 +6203,6 @@ "version": "10.1.1", "resolved": "https://registry.npmjs.org/minimatch/-/minimatch-10.1.1.tgz", "integrity": "sha512-enIvLvRAFZYXJzkCYG5RKmPfrFArdLv+R+lbQ53BmIMLIry74bjKzX6iHAm8WYamJkhSSEabrWN5D97XnKObjQ==", - "dev": true, "license": "BlueOak-1.0.0", "dependencies": { "@isaacs/brace-expansion": "^5.0.0" @@ -6875,7 +6871,6 @@ "version": "4.1.1", "resolved": "https://registry.npmjs.org/jackspeak/-/jackspeak-4.1.1.tgz", "integrity": "sha512-zptv57P3GpL+O0I7VdMJNBZCu+BPHVQUk55Ft8/QCJjTVxrnJHuVuX/0Bl2A6/+2oyR/ZMEuFKwmzqqZ/U5nPQ==", - "dev": true, "license": "BlueOak-1.0.0", "dependencies": { "@isaacs/cliui": "^8.0.2" @@ -7118,7 +7113,6 @@ "version": "11.1.0", "resolved": "https://registry.npmjs.org/lru-cache/-/lru-cache-11.1.0.tgz", "integrity": "sha512-QIXZUBJUx+2zHUdQujWejBkcD9+cs94tLn0+YL8UrCh+D5sCXZ4c7LaEH48pNwRY3MLDgqUFyhlCyjJPf1WP0A==", - "dev": true, "license": "ISC", "engines": { "node": "20 || >=22" @@ -7776,7 +7770,6 @@ "version": "2.0.0", "resolved": "https://registry.npmjs.org/path-scurry/-/path-scurry-2.0.0.tgz", "integrity": "sha512-ypGJsmGtdXUOeM5u93TyeIEfEhM6s+ljAhrk5vAvSx8uyY/02OvrZnA0YNGUrPXfpJMgI1ODd3nwz8Npx4O4cg==", - "dev": true, "license": "BlueOak-1.0.0", "dependencies": { "lru-cache": "^11.0.0", diff --git a/src/feature-flags.ts b/src/feature-flags.ts index 133496979..10e2e296c 100644 --- a/src/feature-flags.ts +++ b/src/feature-flags.ts @@ -77,6 +77,7 @@ export enum Feature { OverlayAnalysisSwift = "overlay_analysis_swift", PythonDefaultIsToNotExtractStdlib = "python_default_is_to_not_extract_stdlib", QaTelemetryEnabled = "qa_telemetry_enabled", + UploadOverlayDbToApi = "upload_overlay_db_to_api", UseRepositoryProperties = "use_repository_properties", ValidateDbConfig = "validate_db_config", } @@ -166,6 +167,11 @@ export const featureConfig: Record< legacyApi: true, minimumVersion: undefined, }, + [Feature.JavaMinimizeDependencyJars]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0", + }, [Feature.OverlayAnalysis]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -277,21 +283,21 @@ export const featureConfig: Record< minimumVersion: undefined, toolsFeature: ToolsFeature.PythonDefaultIsToNotExtractStdlib, }, - [Feature.UseRepositoryProperties]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: undefined, - }, [Feature.QaTelemetryEnabled]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: undefined, }, - [Feature.JavaMinimizeDependencyJars]: { + [Feature.UploadOverlayDbToApi]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0", + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: undefined, + }, + [Feature.UseRepositoryProperties]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: undefined, }, [Feature.ValidateDbConfig]: { defaultValue: false, From 31042e9879e337440768edf221e69d5e7d21a3b1 Mon Sep 17 00:00:00 2001 From: Henry Mercer Date: Tue, 18 Nov 2025 15:27:34 +0000 Subject: [PATCH 38/51] Rename function calls to make destructive operation clearer --- lib/analyze-action.js | 14 ++++++++++---- src/analyze-action.ts | 20 ++++++++++++++------ src/database-upload.test.ts | 16 ++++++++-------- src/database-upload.ts | 2 +- src/overlay-database-utils.ts | 2 +- 5 files changed, 34 insertions(+), 20 deletions(-) diff --git a/lib/analyze-action.js b/lib/analyze-action.js index a2fe4e0b0..389d3e56e 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -88525,7 +88525,7 @@ function checkOverlayBaseDatabase(config, logger, warningPrefix) { } return true; } -async function uploadOverlayBaseDatabaseToCache(codeql, config, logger) { +async function cleanupAndUploadOverlayBaseDatabaseToCache(codeql, config, logger) { const overlayDatabaseMode = config.overlayDatabaseMode; if (overlayDatabaseMode !== "overlay-base" /* OverlayBase */) { logger.debug( @@ -91672,7 +91672,7 @@ async function warnIfGoInstalledAfterInit(config, logger) { // src/database-upload.ts var fs13 = __toESM(require("fs")); -async function uploadDatabases(repositoryNwo, codeql, config, apiDetails, logger) { +async function cleanupAndUploadDatabases(repositoryNwo, codeql, config, apiDetails, logger) { if (getRequiredInput("upload-database") !== "true") { logger.debug("Database upload disabled in workflow. Skipping upload."); return; @@ -94053,8 +94053,14 @@ async function run() { } else { logger.info("Not uploading results"); } - await uploadOverlayBaseDatabaseToCache(codeql, config, logger); - await uploadDatabases(repositoryNwo, codeql, config, apiDetails, logger); + await cleanupAndUploadOverlayBaseDatabaseToCache(codeql, config, logger); + await cleanupAndUploadDatabases( + repositoryNwo, + codeql, + config, + apiDetails, + logger + ); const trapCacheUploadStartTime = import_perf_hooks3.performance.now(); didUploadTrapCaches = await uploadTrapCaches(codeql, config, logger); trapCacheUploadTime = import_perf_hooks3.performance.now() - trapCacheUploadStartTime; diff --git a/src/analyze-action.ts b/src/analyze-action.ts index 3ab1dd132..a101a0187 100644 --- a/src/analyze-action.ts +++ b/src/analyze-action.ts @@ -25,7 +25,7 @@ import { isCodeQualityEnabled, isCodeScanningEnabled, } from "./config-utils"; -import { uploadDatabases } from "./database-upload"; +import { cleanupAndUploadDatabases } from "./database-upload"; import { DependencyCacheUploadStatusReport, uploadDependencyCaches, @@ -35,7 +35,7 @@ import { EnvVar } from "./environment"; import { Feature, Features } from "./feature-flags"; import { KnownLanguage } from "./languages"; import { getActionsLogger, Logger } from "./logging"; -import { uploadOverlayBaseDatabaseToCache } from "./overlay-database-utils"; +import { cleanupAndUploadOverlayBaseDatabaseToCache } from "./overlay-database-utils"; import { getRepositoryNwo } from "./repository"; import * as statusReport from "./status-report"; import { @@ -417,12 +417,20 @@ async function run() { } // Possibly upload the overlay-base database to actions cache. - // If databases are to be uploaded, they will first be cleaned up at the overlay level. - await uploadOverlayBaseDatabaseToCache(codeql, config, logger); + // Note: Take care with the ordering of this call since databases may be cleaned up + // at the `overlay` level. + await cleanupAndUploadOverlayBaseDatabaseToCache(codeql, config, logger); // Possibly upload the database bundles for remote queries. - // If databases are to be uploaded, they will first be cleaned up at the clear level. - await uploadDatabases(repositoryNwo, codeql, config, apiDetails, logger); + // Note: Take care with the ordering of this call since databases may be cleaned up + // at the `overlay` or `clear` level. + await cleanupAndUploadDatabases( + repositoryNwo, + codeql, + config, + apiDetails, + logger, + ); // Possibly upload the TRAP caches for later re-use const trapCacheUploadStartTime = performance.now(); diff --git a/src/database-upload.test.ts b/src/database-upload.test.ts index 6c986fb7f..92f3d75b5 100644 --- a/src/database-upload.test.ts +++ b/src/database-upload.test.ts @@ -10,7 +10,7 @@ import { GitHubApiDetails } from "./api-client"; import * as apiClient from "./api-client"; import { createStubCodeQL } from "./codeql"; import { Config } from "./config-utils"; -import { uploadDatabases } from "./database-upload"; +import { cleanupAndUploadDatabases } from "./database-upload"; import * as gitUtils from "./git-utils"; import { KnownLanguage } from "./languages"; import { RepositoryNwo } from "./repository"; @@ -91,7 +91,7 @@ test("Abort database upload if 'upload-database' input set to false", async (t) sinon.stub(gitUtils, "isAnalyzingDefaultBranch").resolves(true); const loggedMessages = []; - await uploadDatabases( + await cleanupAndUploadDatabases( testRepoName, getCodeQL(), getTestConfig(tmpDir), @@ -121,7 +121,7 @@ test("Abort database upload if 'analysis-kinds: code-scanning' is not enabled", await mockHttpRequests(201); const loggedMessages = []; - await uploadDatabases( + await cleanupAndUploadDatabases( testRepoName, getCodeQL(), { @@ -155,7 +155,7 @@ test("Abort database upload if running against GHES", async (t) => { config.gitHubVersion = { type: GitHubVariant.GHES, version: "3.0" }; const loggedMessages = []; - await uploadDatabases( + await cleanupAndUploadDatabases( testRepoName, getCodeQL(), config, @@ -183,7 +183,7 @@ test("Abort database upload if not analyzing default branch", async (t) => { sinon.stub(gitUtils, "isAnalyzingDefaultBranch").resolves(false); const loggedMessages = []; - await uploadDatabases( + await cleanupAndUploadDatabases( testRepoName, getCodeQL(), getTestConfig(tmpDir), @@ -212,7 +212,7 @@ test("Don't crash if uploading a database fails", async (t) => { await mockHttpRequests(500); const loggedMessages = [] as LoggedMessage[]; - await uploadDatabases( + await cleanupAndUploadDatabases( testRepoName, getCodeQL(), getTestConfig(tmpDir), @@ -243,7 +243,7 @@ test("Successfully uploading a database to github.com", async (t) => { await mockHttpRequests(201); const loggedMessages = [] as LoggedMessage[]; - await uploadDatabases( + await cleanupAndUploadDatabases( testRepoName, getCodeQL(), getTestConfig(tmpDir), @@ -272,7 +272,7 @@ test("Successfully uploading a database to GHEC-DR", async (t) => { const databaseUploadSpy = await mockHttpRequests(201); const loggedMessages = [] as LoggedMessage[]; - await uploadDatabases( + await cleanupAndUploadDatabases( testRepoName, getCodeQL(), getTestConfig(tmpDir), diff --git a/src/database-upload.ts b/src/database-upload.ts index 69175178c..7a147fbc7 100644 --- a/src/database-upload.ts +++ b/src/database-upload.ts @@ -11,7 +11,7 @@ import { RepositoryNwo } from "./repository"; import * as util from "./util"; import { bundleDb, parseGitHubUrl } from "./util"; -export async function uploadDatabases( +export async function cleanupAndUploadDatabases( repositoryNwo: RepositoryNwo, codeql: CodeQL, config: Config, diff --git a/src/overlay-database-utils.ts b/src/overlay-database-utils.ts index ebb020ba8..5b0b4643d 100644 --- a/src/overlay-database-utils.ts +++ b/src/overlay-database-utils.ts @@ -204,7 +204,7 @@ export function checkOverlayBaseDatabase( * @returns A promise that resolves to true if the upload was performed and * successfully completed, or false otherwise */ -export async function uploadOverlayBaseDatabaseToCache( +export async function cleanupAndUploadOverlayBaseDatabaseToCache( codeql: CodeQL, config: Config, logger: Logger, From c649c5993d1fc35396730027271e6afcc384ec39 Mon Sep 17 00:00:00 2001 From: Henry Mercer Date: Tue, 18 Nov 2025 17:38:24 +0000 Subject: [PATCH 39/51] Upload overlay base DB to API behind FF --- lib/analyze-action.js | 8 +++++--- src/analyze-action.ts | 1 + src/codeql.ts | 9 ++++++--- src/database-upload.test.ts | 8 ++++++++ src/database-upload.ts | 13 +++++++++++-- src/overlay-database-utils.ts | 3 ++- src/util.ts | 5 +++++ 7 files changed, 38 insertions(+), 9 deletions(-) diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 389d3e56e..a0fbdc19c 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -88554,7 +88554,7 @@ async function cleanupAndUploadOverlayBaseDatabaseToCache(codeql, config, logger return false; } await withGroupAsync("Cleaning up databases", async () => { - await codeql.databaseCleanupCluster(config, "overlay"); + await codeql.databaseCleanupCluster(config, "overlay" /* Overlay */); }); const dbLocation = config.dbLocation; const databaseSizeBytes = await tryGetFolderBytes(dbLocation, logger); @@ -91672,7 +91672,7 @@ async function warnIfGoInstalledAfterInit(config, logger) { // src/database-upload.ts var fs13 = __toESM(require("fs")); -async function cleanupAndUploadDatabases(repositoryNwo, codeql, config, apiDetails, logger) { +async function cleanupAndUploadDatabases(repositoryNwo, codeql, config, apiDetails, features, logger) { if (getRequiredInput("upload-database") !== "true") { logger.debug("Database upload disabled in workflow. Skipping upload."); return; @@ -91695,8 +91695,9 @@ async function cleanupAndUploadDatabases(repositoryNwo, codeql, config, apiDetai logger.debug("Not analyzing default branch. Skipping upload."); return; } + const cleanupLevel = config.overlayDatabaseMode === "overlay-base" /* OverlayBase */ && await features.getValue("upload_overlay_db_to_api" /* UploadOverlayDbToApi */) ? "overlay" /* Overlay */ : "clear" /* Clear */; await withGroupAsync("Cleaning up databases", async () => { - await codeql.databaseCleanupCluster(config, "clear"); + await codeql.databaseCleanupCluster(config, cleanupLevel); }); const client = getApiClient(); const uploadsUrl = new URL(parseGitHubUrl(apiDetails.url)); @@ -94059,6 +94060,7 @@ async function run() { codeql, config, apiDetails, + features, logger ); const trapCacheUploadStartTime = import_perf_hooks3.performance.now(); diff --git a/src/analyze-action.ts b/src/analyze-action.ts index a101a0187..abbf23972 100644 --- a/src/analyze-action.ts +++ b/src/analyze-action.ts @@ -429,6 +429,7 @@ async function run() { codeql, config, apiDetails, + features, logger, ); diff --git a/src/codeql.ts b/src/codeql.ts index bc1d00401..fbb55dbea 100644 --- a/src/codeql.ts +++ b/src/codeql.ts @@ -35,7 +35,7 @@ import { ToolsDownloadStatusReport } from "./tools-download"; import { ToolsFeature, isSupportedToolsFeature } from "./tools-features"; import { shouldEnableIndirectTracing } from "./tracer-config"; import * as util from "./util"; -import { BuildMode, getErrorMessage } from "./util"; +import { BuildMode, CleanupLevel, getErrorMessage } from "./util"; type Options = Array; @@ -141,7 +141,10 @@ export interface CodeQL { /** * Clean up all the databases within a database cluster. */ - databaseCleanupCluster(config: Config, cleanupLevel: string): Promise; + databaseCleanupCluster( + config: Config, + cleanupLevel: CleanupLevel, + ): Promise; /** * Run 'codeql database bundle'. */ @@ -878,7 +881,7 @@ export async function getCodeQLForCmd( }, async databaseCleanupCluster( config: Config, - cleanupLevel: string, + cleanupLevel: CleanupLevel, ): Promise { const cacheCleanupFlag = (await util.codeQlVersionAtLeast( this, diff --git a/src/database-upload.test.ts b/src/database-upload.test.ts index 92f3d75b5..e07ff1da2 100644 --- a/src/database-upload.test.ts +++ b/src/database-upload.test.ts @@ -15,6 +15,7 @@ import * as gitUtils from "./git-utils"; import { KnownLanguage } from "./languages"; import { RepositoryNwo } from "./repository"; import { + createFeatures, createTestConfig, getRecordingLogger, LoggedMessage, @@ -96,6 +97,7 @@ test("Abort database upload if 'upload-database' input set to false", async (t) getCodeQL(), getTestConfig(tmpDir), testApiDetails, + createFeatures([]), getRecordingLogger(loggedMessages), ); t.assert( @@ -129,6 +131,7 @@ test("Abort database upload if 'analysis-kinds: code-scanning' is not enabled", analysisKinds: [AnalysisKind.CodeQuality], }, testApiDetails, + createFeatures([]), getRecordingLogger(loggedMessages), ); t.assert( @@ -160,6 +163,7 @@ test("Abort database upload if running against GHES", async (t) => { getCodeQL(), config, testApiDetails, + createFeatures([]), getRecordingLogger(loggedMessages), ); t.assert( @@ -188,6 +192,7 @@ test("Abort database upload if not analyzing default branch", async (t) => { getCodeQL(), getTestConfig(tmpDir), testApiDetails, + createFeatures([]), getRecordingLogger(loggedMessages), ); t.assert( @@ -217,6 +222,7 @@ test("Don't crash if uploading a database fails", async (t) => { getCodeQL(), getTestConfig(tmpDir), testApiDetails, + createFeatures([]), getRecordingLogger(loggedMessages), ); @@ -248,6 +254,7 @@ test("Successfully uploading a database to github.com", async (t) => { getCodeQL(), getTestConfig(tmpDir), testApiDetails, + createFeatures([]), getRecordingLogger(loggedMessages), ); t.assert( @@ -281,6 +288,7 @@ test("Successfully uploading a database to GHEC-DR", async (t) => { url: "https://tenant.ghe.com", apiURL: undefined, }, + createFeatures([]), getRecordingLogger(loggedMessages), ); t.assert( diff --git a/src/database-upload.ts b/src/database-upload.ts index 7a147fbc7..d99df14c3 100644 --- a/src/database-upload.ts +++ b/src/database-upload.ts @@ -5,17 +5,20 @@ import { AnalysisKind } from "./analyses"; import { getApiClient, GitHubApiDetails } from "./api-client"; import { type CodeQL } from "./codeql"; import { Config } from "./config-utils"; +import { Feature, FeatureEnablement } from "./feature-flags"; import * as gitUtils from "./git-utils"; import { Logger, withGroupAsync } from "./logging"; +import { OverlayDatabaseMode } from "./overlay-database-utils"; import { RepositoryNwo } from "./repository"; import * as util from "./util"; -import { bundleDb, parseGitHubUrl } from "./util"; +import { bundleDb, CleanupLevel, parseGitHubUrl } from "./util"; export async function cleanupAndUploadDatabases( repositoryNwo: RepositoryNwo, codeql: CodeQL, config: Config, apiDetails: GitHubApiDetails, + features: FeatureEnablement, logger: Logger, ): Promise { if (actionsUtil.getRequiredInput("upload-database") !== "true") { @@ -50,10 +53,16 @@ export async function cleanupAndUploadDatabases( return; } + const cleanupLevel = + config.overlayDatabaseMode === OverlayDatabaseMode.OverlayBase && + (await features.getValue(Feature.UploadOverlayDbToApi)) + ? CleanupLevel.Overlay + : CleanupLevel.Clear; + // Clean up the database, since intermediate results may still be written to the // database if there is high RAM pressure. await withGroupAsync("Cleaning up databases", async () => { - await codeql.databaseCleanupCluster(config, "clear"); + await codeql.databaseCleanupCluster(config, cleanupLevel); }); const client = getApiClient(); diff --git a/src/overlay-database-utils.ts b/src/overlay-database-utils.ts index 5b0b4643d..89056cce0 100644 --- a/src/overlay-database-utils.ts +++ b/src/overlay-database-utils.ts @@ -16,6 +16,7 @@ import { type Config } from "./config-utils"; import { getCommitOid, getFileOidsUnderPath } from "./git-utils"; import { Logger, withGroupAsync } from "./logging"; import { + CleanupLevel, getErrorMessage, isInTestMode, tryGetFolderBytes, @@ -242,7 +243,7 @@ export async function cleanupAndUploadOverlayBaseDatabaseToCache( // Clean up the database using the overlay cleanup level. await withGroupAsync("Cleaning up databases", async () => { - await codeql.databaseCleanupCluster(config, "overlay"); + await codeql.databaseCleanupCluster(config, CleanupLevel.Overlay); }); const dbLocation = config.dbLocation; diff --git a/src/util.ts b/src/util.ts index f23c3be7d..546f328a9 100644 --- a/src/util.ts +++ b/src/util.ts @@ -1314,3 +1314,8 @@ export function unsafeEntriesInvariant>( ([_, val]) => val !== undefined, ) as Array<[keyof T, Exclude]>; } + +export enum CleanupLevel { + Clear = "clear", + Overlay = "overlay", +} From ed80d6e5e959958d907c93c7e350411df87b57aa Mon Sep 17 00:00:00 2001 From: Kasper Svendsen Date: Wed, 19 Nov 2025 07:54:05 +0100 Subject: [PATCH 40/51] Overlay: Reorder available disk space check --- lib/init-action.js | 14 +++++++------- src/config-utils.ts | 18 +++++++++--------- 2 files changed, 16 insertions(+), 16 deletions(-) diff --git a/lib/init-action.js b/lib/init-action.js index ea9246bb5..dc49463d7 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -86920,7 +86920,12 @@ async function getOverlayDatabaseMode(codeql, features, languages, sourceRoot, b logger.info( `Setting overlay database mode to ${overlayDatabaseMode} from the CODEQL_OVERLAY_DATABASE_MODE environment variable.` ); - } else { + } else if (await isOverlayAnalysisFeatureEnabled( + features, + codeql, + languages, + codeScanningConfig + )) { const diskUsage = await checkDiskUsage(logger); if (diskUsage === void 0 || diskUsage.numAvailableBytes < OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES) { const diskSpaceMb = diskUsage === void 0 ? 0 : diskUsage.numAvailableBytes / 1e6; @@ -86929,12 +86934,7 @@ async function getOverlayDatabaseMode(codeql, features, languages, sourceRoot, b logger.info( `Setting overlay database mode to ${overlayDatabaseMode} due to insufficient disk space (${diskSpaceMb} MB).` ); - } else if (await isOverlayAnalysisFeatureEnabled( - features, - codeql, - languages, - codeScanningConfig - )) { + } else { if (isAnalyzingPullRequest()) { overlayDatabaseMode = "overlay" /* Overlay */; useOverlayDatabaseCaching = true; diff --git a/src/config-utils.ts b/src/config-utils.ts index f910e5d90..03d88722f 100644 --- a/src/config-utils.ts +++ b/src/config-utils.ts @@ -679,7 +679,14 @@ export async function getOverlayDatabaseMode( `Setting overlay database mode to ${overlayDatabaseMode} ` + "from the CODEQL_OVERLAY_DATABASE_MODE environment variable.", ); - } else { + } else if ( + await isOverlayAnalysisFeatureEnabled( + features, + codeql, + languages, + codeScanningConfig, + ) + ) { const diskUsage = await checkDiskUsage(logger); if ( diskUsage === undefined || @@ -693,14 +700,7 @@ export async function getOverlayDatabaseMode( `Setting overlay database mode to ${overlayDatabaseMode} ` + `due to insufficient disk space (${diskSpaceMb} MB).`, ); - } else if ( - await isOverlayAnalysisFeatureEnabled( - features, - codeql, - languages, - codeScanningConfig, - ) - ) { + } else { if (isAnalyzingPullRequest()) { overlayDatabaseMode = OverlayDatabaseMode.Overlay; useOverlayDatabaseCaching = true; From 4eccb3798e6657749ccc8a7b4813a58748ec16d5 Mon Sep 17 00:00:00 2001 From: Kasper Svendsen Date: Wed, 19 Nov 2025 07:59:59 +0100 Subject: [PATCH 41/51] Overlay: Round available disk space in MB --- lib/init-action.js | 2 +- src/config-utils.ts | 4 +++- 2 files changed, 4 insertions(+), 2 deletions(-) diff --git a/lib/init-action.js b/lib/init-action.js index dc49463d7..ff9b72333 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -86928,7 +86928,7 @@ async function getOverlayDatabaseMode(codeql, features, languages, sourceRoot, b )) { const diskUsage = await checkDiskUsage(logger); if (diskUsage === void 0 || diskUsage.numAvailableBytes < OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES) { - const diskSpaceMb = diskUsage === void 0 ? 0 : diskUsage.numAvailableBytes / 1e6; + const diskSpaceMb = diskUsage === void 0 ? 0 : Math.round(diskUsage.numAvailableBytes / 1e6); overlayDatabaseMode = "none" /* None */; useOverlayDatabaseCaching = false; logger.info( diff --git a/src/config-utils.ts b/src/config-utils.ts index 03d88722f..257257344 100644 --- a/src/config-utils.ts +++ b/src/config-utils.ts @@ -693,7 +693,9 @@ export async function getOverlayDatabaseMode( diskUsage.numAvailableBytes < OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES ) { const diskSpaceMb = - diskUsage === undefined ? 0 : diskUsage.numAvailableBytes / 1_000_000; + diskUsage === undefined + ? 0 + : Math.round(diskUsage.numAvailableBytes / 1_000_000); overlayDatabaseMode = OverlayDatabaseMode.None; useOverlayDatabaseCaching = false; logger.info( From de74d762a3a5ec4440385fdd84b1dbe329b2cbf8 Mon Sep 17 00:00:00 2001 From: Kasper Svendsen Date: Wed, 19 Nov 2025 13:02:54 +0100 Subject: [PATCH 42/51] Overlay: Increase minimum CLI version --- lib/analyze-action-post.js | 2 +- lib/analyze-action.js | 2 +- lib/autobuild-action.js | 2 +- lib/init-action-post.js | 2 +- lib/init-action.js | 2 +- lib/resolve-environment-action.js | 2 +- lib/setup-codeql-action.js | 2 +- lib/start-proxy-action-post.js | 2 +- lib/start-proxy-action.js | 2 +- lib/upload-lib.js | 2 +- lib/upload-sarif-action-post.js | 2 +- lib/upload-sarif-action.js | 2 +- src/overlay-database-utils.ts | 2 +- 13 files changed, 13 insertions(+), 13 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index 21e183cb9..33a78937f 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -119948,7 +119948,7 @@ function withGroup(groupName, f) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 4f2ba5b16..a0f5189d2 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -88453,7 +88453,7 @@ function formatDuration(durationMs) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index ce7434945..e246d9418 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -83890,7 +83890,7 @@ function getActionsLogger() { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { diff --git a/lib/init-action-post.js b/lib/init-action-post.js index c7dcc67c4..7c6fe55ff 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -123326,7 +123326,7 @@ function formatDuration(durationMs) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { diff --git a/lib/init-action.js b/lib/init-action.js index 2c18eb561..7f32a6b6c 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -85855,7 +85855,7 @@ function formatDuration(durationMs) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index a8c741668..22721b588 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -83883,7 +83883,7 @@ function getActionsLogger() { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index 89ff54951..1a1fa9c21 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -83792,7 +83792,7 @@ function formatDuration(durationMs) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index a44057c79..2199b9993 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -119417,7 +119417,7 @@ function getActionsLogger() { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index ce895e450..694eb371b 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -99970,7 +99970,7 @@ async function getRef() { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; diff --git a/lib/upload-lib.js b/lib/upload-lib.js index 79a70e3d7..1845f1ac2 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -86948,7 +86948,7 @@ function formatDuration(durationMs) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index 4462a6828..88c20f390 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -119579,7 +119579,7 @@ function withGroup(groupName, f) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 84ffad340..95c7fd492 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -86742,7 +86742,7 @@ function formatDuration(durationMs) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { diff --git a/src/overlay-database-utils.ts b/src/overlay-database-utils.ts index 89056cce0..e0a43391b 100644 --- a/src/overlay-database-utils.ts +++ b/src/overlay-database-utils.ts @@ -29,7 +29,7 @@ export enum OverlayDatabaseMode { None = "none", } -export const CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +export const CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; /** * The maximum (uncompressed) size of the overlay base database that we will From ac359aad20e59fd46ecd05e63c6d4b99cad25272 Mon Sep 17 00:00:00 2001 From: Henry Mercer Date: Wed, 19 Nov 2025 14:59:16 +0000 Subject: [PATCH 43/51] Add return type --- src/setup-codeql.ts | 10 +++++++++- 1 file changed, 9 insertions(+), 1 deletion(-) diff --git a/src/setup-codeql.ts b/src/setup-codeql.ts index 6a412d1df..16375421a 100644 --- a/src/setup-codeql.ts +++ b/src/setup-codeql.ts @@ -718,6 +718,14 @@ function getCanonicalToolcacheVersion( return cliVersion; } +interface SetupCodeQLResult { + codeqlFolder: string; + toolsDownloadStatusReport?: ToolsDownloadStatusReport; + toolsSource: ToolsSource; + toolsVersion: string; + zstdAvailability: tar.ZstdAvailability; +} + /** * Obtains the CodeQL bundle, installs it in the toolcache if appropriate, and extracts it. * @@ -731,7 +739,7 @@ export async function setupCodeQLBundle( defaultCliVersion: CodeQLDefaultVersionInfo, features: FeatureEnablement, logger: Logger, -) { +): Promise { if (!(await util.isBinaryAccessible("tar", logger))) { throw new util.ConfigurationError( "Could not find tar in PATH, so unable to extract CodeQL bundle.", From 1d2a238d7d52b563f78fb3bf1deebc2b18d98eb1 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 24 Nov 2025 07:51:46 +0000 Subject: [PATCH 44/51] Update default bundle to codeql-bundle-v2.23.6 --- lib/analyze-action.js | 4 ++-- lib/autobuild-action.js | 4 ++-- lib/defaults.json | 8 ++++---- lib/init-action-post.js | 4 ++-- lib/init-action.js | 4 ++-- lib/setup-codeql-action.js | 4 ++-- lib/start-proxy-action.js | 4 ++-- lib/upload-lib.js | 4 ++-- lib/upload-sarif-action.js | 4 ++-- src/defaults.json | 8 ++++---- 10 files changed, 24 insertions(+), 24 deletions(-) diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 5781e6941..6f1a1bf42 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -88211,8 +88211,8 @@ var path4 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.5"; -var cliVersion = "2.23.5"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index f286a07cf..9a7251809 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -83701,8 +83701,8 @@ var path3 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.5"; -var cliVersion = "2.23.5"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/overlay-database-utils.ts var fs2 = __toESM(require("fs")); diff --git a/lib/defaults.json b/lib/defaults.json index 9be5b5476..835b6a33b 100644 --- a/lib/defaults.json +++ b/lib/defaults.json @@ -1,6 +1,6 @@ { - "bundleVersion": "codeql-bundle-v2.23.5", - "cliVersion": "2.23.5", - "priorBundleVersion": "codeql-bundle-v2.23.3", - "priorCliVersion": "2.23.3" + "bundleVersion": "codeql-bundle-v2.23.6", + "cliVersion": "2.23.6", + "priorBundleVersion": "codeql-bundle-v2.23.5", + "priorCliVersion": "2.23.5" } diff --git a/lib/init-action-post.js b/lib/init-action-post.js index fdc23d247..89948b8a1 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -123084,8 +123084,8 @@ var path4 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.5"; -var cliVersion = "2.23.5"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); diff --git a/lib/init-action.js b/lib/init-action.js index 6f826febd..f8407c208 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -85635,8 +85635,8 @@ var path5 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.5"; -var cliVersion = "2.23.5"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index cad0195ad..f1182b65c 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -83589,8 +83589,8 @@ var path4 = __toESM(require("path")); var semver3 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.5"; -var cliVersion = "2.23.5"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index 3693c9670..3c2490783 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -99684,8 +99684,8 @@ function getActionsLogger() { var core7 = __toESM(require_core()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.5"; -var cliVersion = "2.23.5"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/languages.ts var KnownLanguage = /* @__PURE__ */ ((KnownLanguage2) => { diff --git a/lib/upload-lib.js b/lib/upload-lib.js index 938245c21..53eaa204e 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -86724,8 +86724,8 @@ var path4 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.5"; -var cliVersion = "2.23.5"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 79778f00d..574910f02 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -86505,8 +86505,8 @@ var path4 = __toESM(require("path")); var semver3 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.5"; -var cliVersion = "2.23.5"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); diff --git a/src/defaults.json b/src/defaults.json index 9be5b5476..835b6a33b 100644 --- a/src/defaults.json +++ b/src/defaults.json @@ -1,6 +1,6 @@ { - "bundleVersion": "codeql-bundle-v2.23.5", - "cliVersion": "2.23.5", - "priorBundleVersion": "codeql-bundle-v2.23.3", - "priorCliVersion": "2.23.3" + "bundleVersion": "codeql-bundle-v2.23.6", + "cliVersion": "2.23.6", + "priorBundleVersion": "codeql-bundle-v2.23.5", + "priorCliVersion": "2.23.5" } From ecc87875ee10fd563cebc295e45bea8312e2ce49 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 24 Nov 2025 07:51:53 +0000 Subject: [PATCH 45/51] Add changelog note --- CHANGELOG.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 3ec876d4e..f637c0040 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,7 +4,7 @@ See the [releases page](https://github.com/github/codeql-action/releases) for th ## [UNRELEASED] -No user facing changes. +- Update default CodeQL bundle version to 2.23.6. [#3321](https://github.com/github/codeql-action/pull/3321) ## 4.31.4 - 18 Nov 2025 From 81f6d649ae64626b3035526b0389bfa8802b6df3 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 24 Nov 2025 09:03:58 +0000 Subject: [PATCH 46/51] Update changelog for v4.31.5 --- CHANGELOG.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index f637c0040..762aa1db8 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -2,7 +2,7 @@ See the [releases page](https://github.com/github/codeql-action/releases) for the relevant changes to the CodeQL CLI and language packs. -## [UNRELEASED] +## 4.31.5 - 24 Nov 2025 - Update default CodeQL bundle version to 2.23.6. [#3321](https://github.com/github/codeql-action/pull/3321) From 29e11fdce1ae617d40492467e777f61f2d9fc0c0 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 24 Nov 2025 09:31:18 +0000 Subject: [PATCH 47/51] Update changelog and version after v4.31.5 --- CHANGELOG.md | 4 ++++ package-lock.json | 4 ++-- package.json | 2 +- 3 files changed, 7 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 762aa1db8..1359cdfd9 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -2,6 +2,10 @@ See the [releases page](https://github.com/github/codeql-action/releases) for the relevant changes to the CodeQL CLI and language packs. +## [UNRELEASED] + +No user facing changes. + ## 4.31.5 - 24 Nov 2025 - Update default CodeQL bundle version to 2.23.6. [#3321](https://github.com/github/codeql-action/pull/3321) diff --git a/package-lock.json b/package-lock.json index 3ee4a5b89..4c6ca8624 100644 --- a/package-lock.json +++ b/package-lock.json @@ -1,12 +1,12 @@ { "name": "codeql", - "version": "4.31.5", + "version": "4.31.6", "lockfileVersion": 3, "requires": true, "packages": { "": { "name": "codeql", - "version": "4.31.5", + "version": "4.31.6", "license": "MIT", "dependencies": { "@actions/artifact": "^4.0.0", diff --git a/package.json b/package.json index 61317b90a..22d581776 100644 --- a/package.json +++ b/package.json @@ -1,6 +1,6 @@ { "name": "codeql", - "version": "4.31.5", + "version": "4.31.6", "private": true, "description": "CodeQL action", "scripts": { From 478350182f3269d74025e346d386b05203bda49f Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 24 Nov 2025 10:55:14 +0000 Subject: [PATCH 48/51] Rebuild --- lib/analyze-action-post.js | 2 +- lib/analyze-action.js | 2 +- lib/autobuild-action.js | 2 +- lib/init-action-post.js | 2 +- lib/init-action.js | 2 +- lib/resolve-environment-action.js | 2 +- lib/setup-codeql-action.js | 2 +- lib/start-proxy-action-post.js | 2 +- lib/start-proxy-action.js | 2 +- lib/upload-lib.js | 2 +- lib/upload-sarif-action-post.js | 2 +- lib/upload-sarif-action.js | 2 +- 12 files changed, 12 insertions(+), 12 deletions(-) diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index 13589f496..37725d00b 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.5", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { diff --git a/lib/analyze-action.js b/lib/analyze-action.js index 6f1a1bf42..8206afce0 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.5", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index 9a7251809..d4b7fc1f6 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.5", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { diff --git a/lib/init-action-post.js b/lib/init-action-post.js index 89948b8a1..62e78df8a 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.5", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { diff --git a/lib/init-action.js b/lib/init-action.js index f8407c208..185510e02 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.5", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index 48ebce48f..cd65a4bf1 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.5", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index f1182b65c..780d2cc6d 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.5", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index cdac66bef..c78e8262a 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.5", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index 3c2490783..f0d7eb571 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -47285,7 +47285,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.5", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { diff --git a/lib/upload-lib.js b/lib/upload-lib.js index 53eaa204e..de44834ac 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -28924,7 +28924,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.5", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index 87ef62a45..f95b705fa 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.5", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 574910f02..f8ea28a2d 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.5", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { From 6feac2b36a5ca1b9bef24d689424860a700aaf65 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 24 Nov 2025 17:59:04 +0000 Subject: [PATCH 49/51] Bump actions/create-github-app-token Bumps the actions-minor group with 1 update in the /.github/workflows directory: [actions/create-github-app-token](https://github.com/actions/create-github-app-token). Updates `actions/create-github-app-token` from 2.1.4 to 2.2.0 - [Release notes](https://github.com/actions/create-github-app-token/releases) - [Commits](https://github.com/actions/create-github-app-token/compare/v2.1.4...v2.2.0) --- updated-dependencies: - dependency-name: actions/create-github-app-token dependency-version: 2.2.0 dependency-type: direct:production update-type: version-update:semver-minor dependency-group: actions-minor ... Signed-off-by: dependabot[bot] --- .github/workflows/post-release-mergeback.yml | 2 +- .github/workflows/rollback-release.yml | 2 +- .github/workflows/update-release-branch.yml | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.github/workflows/post-release-mergeback.yml b/.github/workflows/post-release-mergeback.yml index 1731a78ff..4be56da52 100644 --- a/.github/workflows/post-release-mergeback.yml +++ b/.github/workflows/post-release-mergeback.yml @@ -142,7 +142,7 @@ jobs: token: "${{ secrets.GITHUB_TOKEN }}" - name: Generate token - uses: actions/create-github-app-token@v2.1.4 + uses: actions/create-github-app-token@v2.2.0 id: app-token with: app-id: ${{ vars.AUTOMATION_APP_ID }} diff --git a/.github/workflows/rollback-release.yml b/.github/workflows/rollback-release.yml index 8d8e872fa..a218fd57e 100644 --- a/.github/workflows/rollback-release.yml +++ b/.github/workflows/rollback-release.yml @@ -137,7 +137,7 @@ jobs: - name: Generate token if: github.event_name == 'workflow_dispatch' - uses: actions/create-github-app-token@v2.1.4 + uses: actions/create-github-app-token@v2.2.0 id: app-token with: app-id: ${{ vars.AUTOMATION_APP_ID }} diff --git a/.github/workflows/update-release-branch.yml b/.github/workflows/update-release-branch.yml index 830ed7c2a..74349965b 100644 --- a/.github/workflows/update-release-branch.yml +++ b/.github/workflows/update-release-branch.yml @@ -93,7 +93,7 @@ jobs: pull-requests: write # needed to create pull request steps: - name: Generate token - uses: actions/create-github-app-token@v2.1.4 + uses: actions/create-github-app-token@v2.2.0 id: app-token with: app-id: ${{ vars.AUTOMATION_APP_ID }} From 5bd8069afb7ffe286094a9d3f1026925d4ac7990 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 24 Nov 2025 18:01:10 +0000 Subject: [PATCH 50/51] Bump actions/checkout from 5 to 6 in /.github/workflows Bumps [actions/checkout](https://github.com/actions/checkout) from 5 to 6. - [Release notes](https://github.com/actions/checkout/releases) - [Changelog](https://github.com/actions/checkout/blob/main/CHANGELOG.md) - [Commits](https://github.com/actions/checkout/compare/v5...v6) --- updated-dependencies: - dependency-name: actions/checkout dependency-version: '6' dependency-type: direct:production update-type: version-update:semver-major ... Signed-off-by: dependabot[bot] --- .github/workflows/__all-platform-bundle.yml | 2 +- .github/workflows/__analyze-ref-input.yml | 2 +- .github/workflows/__autobuild-action.yml | 2 +- .../__autobuild-direct-tracing-with-working-dir.yml | 2 +- .github/workflows/__autobuild-working-dir.yml | 2 +- .github/workflows/__build-mode-autobuild.yml | 2 +- .github/workflows/__build-mode-manual.yml | 2 +- .github/workflows/__build-mode-none.yml | 2 +- .github/workflows/__build-mode-rollback.yml | 2 +- .github/workflows/__bundle-from-toolcache.yml | 2 +- .github/workflows/__bundle-toolcache.yml | 2 +- .github/workflows/__bundle-zstd.yml | 2 +- .github/workflows/__cleanup-db-cluster-dir.yml | 2 +- .github/workflows/__config-export.yml | 2 +- .github/workflows/__config-input.yml | 2 +- .github/workflows/__cpp-deptrace-disabled.yml | 2 +- .github/workflows/__cpp-deptrace-enabled-on-macos.yml | 2 +- .github/workflows/__cpp-deptrace-enabled.yml | 2 +- .github/workflows/__diagnostics-export.yml | 2 +- .github/workflows/__export-file-baseline-information.yml | 2 +- .github/workflows/__extractor-ram-threads.yml | 2 +- .github/workflows/__global-proxy.yml | 2 +- .github/workflows/__go-custom-queries.yml | 2 +- .../__go-indirect-tracing-workaround-diagnostic.yml | 2 +- .../__go-indirect-tracing-workaround-no-file-program.yml | 2 +- .github/workflows/__go-indirect-tracing-workaround.yml | 2 +- .github/workflows/__go-tracing-autobuilder.yml | 2 +- .github/workflows/__go-tracing-custom-build-steps.yml | 2 +- .github/workflows/__go-tracing-legacy-workflow.yml | 2 +- .github/workflows/__init-with-registries.yml | 2 +- .github/workflows/__javascript-source-root.yml | 2 +- .github/workflows/__job-run-uuid-sarif.yml | 2 +- .github/workflows/__language-aliases.yml | 2 +- .github/workflows/__local-bundle.yml | 2 +- .github/workflows/__multi-language-autodetect.yml | 2 +- .github/workflows/__overlay-init-fallback.yml | 2 +- .../workflows/__packaging-codescanning-config-inputs-js.yml | 2 +- .github/workflows/__packaging-config-inputs-js.yml | 2 +- .github/workflows/__packaging-config-js.yml | 2 +- .github/workflows/__packaging-inputs-js.yml | 2 +- .github/workflows/__quality-queries.yml | 2 +- .github/workflows/__remote-config.yml | 2 +- .github/workflows/__resolve-environment-action.yml | 2 +- .github/workflows/__rubocop-multi-language.yml | 2 +- .github/workflows/__ruby.yml | 2 +- .github/workflows/__rust.yml | 2 +- .github/workflows/__split-workflow.yml | 2 +- .github/workflows/__start-proxy.yml | 2 +- .github/workflows/__submit-sarif-failure.yml | 4 ++-- .github/workflows/__swift-autobuild.yml | 2 +- .github/workflows/__swift-custom-build.yml | 2 +- .github/workflows/__unset-environment.yml | 2 +- .github/workflows/__upload-ref-sha-input.yml | 2 +- .github/workflows/__upload-sarif.yml | 2 +- .github/workflows/__with-checkout-path.yml | 4 ++-- .github/workflows/check-expected-release-files.yml | 2 +- .github/workflows/codeql.yml | 6 +++--- .github/workflows/codescanning-config-cli.yml | 2 +- .github/workflows/debug-artifacts-failure-safe.yml | 2 +- .github/workflows/debug-artifacts-safe.yml | 2 +- .github/workflows/post-release-mergeback.yml | 2 +- .github/workflows/pr-checks.yml | 6 +++--- .github/workflows/prepare-release.yml | 2 +- .github/workflows/publish-immutable-action.yml | 2 +- .github/workflows/python312-windows.yml | 2 +- .github/workflows/query-filters.yml | 2 +- .github/workflows/rebuild.yml | 2 +- .github/workflows/rollback-release.yml | 2 +- .github/workflows/test-codeql-bundle-all.yml | 2 +- .github/workflows/update-bundle.yml | 2 +- .github/workflows/update-release-branch.yml | 4 ++-- .../update-supported-enterprise-server-versions.yml | 4 ++-- 72 files changed, 80 insertions(+), 80 deletions(-) diff --git a/.github/workflows/__all-platform-bundle.yml b/.github/workflows/__all-platform-bundle.yml index e2b5e69fc..2340be49c 100644 --- a/.github/workflows/__all-platform-bundle.yml +++ b/.github/workflows/__all-platform-bundle.yml @@ -71,7 +71,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__analyze-ref-input.yml b/.github/workflows/__analyze-ref-input.yml index 9efe4a8c3..161942723 100644 --- a/.github/workflows/__analyze-ref-input.yml +++ b/.github/workflows/__analyze-ref-input.yml @@ -77,7 +77,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__autobuild-action.yml b/.github/workflows/__autobuild-action.yml index 0e617afe1..08470bcff 100644 --- a/.github/workflows/__autobuild-action.yml +++ b/.github/workflows/__autobuild-action.yml @@ -61,7 +61,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__autobuild-direct-tracing-with-working-dir.yml b/.github/workflows/__autobuild-direct-tracing-with-working-dir.yml index c1de5c19d..9607fce18 100644 --- a/.github/workflows/__autobuild-direct-tracing-with-working-dir.yml +++ b/.github/workflows/__autobuild-direct-tracing-with-working-dir.yml @@ -63,7 +63,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__autobuild-working-dir.yml b/.github/workflows/__autobuild-working-dir.yml index 3a3ca9e5f..e9d1d7d5d 100644 --- a/.github/workflows/__autobuild-working-dir.yml +++ b/.github/workflows/__autobuild-working-dir.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__build-mode-autobuild.yml b/.github/workflows/__build-mode-autobuild.yml index 878c941a4..87ed95e1e 100644 --- a/.github/workflows/__build-mode-autobuild.yml +++ b/.github/workflows/__build-mode-autobuild.yml @@ -63,7 +63,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__build-mode-manual.yml b/.github/workflows/__build-mode-manual.yml index 4be0c42d1..c164a1a7b 100644 --- a/.github/workflows/__build-mode-manual.yml +++ b/.github/workflows/__build-mode-manual.yml @@ -67,7 +67,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__build-mode-none.yml b/.github/workflows/__build-mode-none.yml index 7584f9065..7bb121810 100644 --- a/.github/workflows/__build-mode-none.yml +++ b/.github/workflows/__build-mode-none.yml @@ -49,7 +49,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__build-mode-rollback.yml b/.github/workflows/__build-mode-rollback.yml index c1f3ccd0c..e9d85968c 100644 --- a/.github/workflows/__build-mode-rollback.yml +++ b/.github/workflows/__build-mode-rollback.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__bundle-from-toolcache.yml b/.github/workflows/__bundle-from-toolcache.yml index 639595af5..96858acd1 100644 --- a/.github/workflows/__bundle-from-toolcache.yml +++ b/.github/workflows/__bundle-from-toolcache.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__bundle-toolcache.yml b/.github/workflows/__bundle-toolcache.yml index de3826b65..59d06b49b 100644 --- a/.github/workflows/__bundle-toolcache.yml +++ b/.github/workflows/__bundle-toolcache.yml @@ -51,7 +51,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__bundle-zstd.yml b/.github/workflows/__bundle-zstd.yml index f5b1ab3aa..18185ada3 100644 --- a/.github/workflows/__bundle-zstd.yml +++ b/.github/workflows/__bundle-zstd.yml @@ -51,7 +51,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__cleanup-db-cluster-dir.yml b/.github/workflows/__cleanup-db-cluster-dir.yml index dfe53c67c..8bf4659ae 100644 --- a/.github/workflows/__cleanup-db-cluster-dir.yml +++ b/.github/workflows/__cleanup-db-cluster-dir.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__config-export.yml b/.github/workflows/__config-export.yml index f01c4ae3d..1c9895854 100644 --- a/.github/workflows/__config-export.yml +++ b/.github/workflows/__config-export.yml @@ -49,7 +49,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__config-input.yml b/.github/workflows/__config-input.yml index 59db10d4d..2a006be21 100644 --- a/.github/workflows/__config-input.yml +++ b/.github/workflows/__config-input.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Install Node.js uses: actions/setup-node@v6 with: diff --git a/.github/workflows/__cpp-deptrace-disabled.yml b/.github/workflows/__cpp-deptrace-disabled.yml index 122159236..2116e5c4f 100644 --- a/.github/workflows/__cpp-deptrace-disabled.yml +++ b/.github/workflows/__cpp-deptrace-disabled.yml @@ -51,7 +51,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__cpp-deptrace-enabled-on-macos.yml b/.github/workflows/__cpp-deptrace-enabled-on-macos.yml index b9669b870..1039cc321 100644 --- a/.github/workflows/__cpp-deptrace-enabled-on-macos.yml +++ b/.github/workflows/__cpp-deptrace-enabled-on-macos.yml @@ -49,7 +49,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__cpp-deptrace-enabled.yml b/.github/workflows/__cpp-deptrace-enabled.yml index bf155a64d..9a57d0041 100644 --- a/.github/workflows/__cpp-deptrace-enabled.yml +++ b/.github/workflows/__cpp-deptrace-enabled.yml @@ -51,7 +51,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__diagnostics-export.yml b/.github/workflows/__diagnostics-export.yml index 9251e04a8..8c05b6d92 100644 --- a/.github/workflows/__diagnostics-export.yml +++ b/.github/workflows/__diagnostics-export.yml @@ -49,7 +49,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__export-file-baseline-information.yml b/.github/workflows/__export-file-baseline-information.yml index 980535c84..7ebf51f3f 100644 --- a/.github/workflows/__export-file-baseline-information.yml +++ b/.github/workflows/__export-file-baseline-information.yml @@ -71,7 +71,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__extractor-ram-threads.yml b/.github/workflows/__extractor-ram-threads.yml index 2d8316f52..09c1cbbf4 100644 --- a/.github/workflows/__extractor-ram-threads.yml +++ b/.github/workflows/__extractor-ram-threads.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__global-proxy.yml b/.github/workflows/__global-proxy.yml index bd5d64b5f..35f1f08fc 100644 --- a/.github/workflows/__global-proxy.yml +++ b/.github/workflows/__global-proxy.yml @@ -61,7 +61,7 @@ jobs: apt install -y gh env: {} - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__go-custom-queries.yml b/.github/workflows/__go-custom-queries.yml index fe35b5b4d..32912ee07 100644 --- a/.github/workflows/__go-custom-queries.yml +++ b/.github/workflows/__go-custom-queries.yml @@ -69,7 +69,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__go-indirect-tracing-workaround-diagnostic.yml b/.github/workflows/__go-indirect-tracing-workaround-diagnostic.yml index 061ad4254..b140b13c4 100644 --- a/.github/workflows/__go-indirect-tracing-workaround-diagnostic.yml +++ b/.github/workflows/__go-indirect-tracing-workaround-diagnostic.yml @@ -57,7 +57,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__go-indirect-tracing-workaround-no-file-program.yml b/.github/workflows/__go-indirect-tracing-workaround-no-file-program.yml index 0a347c65c..d6cf269d7 100644 --- a/.github/workflows/__go-indirect-tracing-workaround-no-file-program.yml +++ b/.github/workflows/__go-indirect-tracing-workaround-no-file-program.yml @@ -57,7 +57,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__go-indirect-tracing-workaround.yml b/.github/workflows/__go-indirect-tracing-workaround.yml index bb811d4d5..8b0c3d13d 100644 --- a/.github/workflows/__go-indirect-tracing-workaround.yml +++ b/.github/workflows/__go-indirect-tracing-workaround.yml @@ -57,7 +57,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__go-tracing-autobuilder.yml b/.github/workflows/__go-tracing-autobuilder.yml index 6d4cc91cc..0d2db3734 100644 --- a/.github/workflows/__go-tracing-autobuilder.yml +++ b/.github/workflows/__go-tracing-autobuilder.yml @@ -91,7 +91,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__go-tracing-custom-build-steps.yml b/.github/workflows/__go-tracing-custom-build-steps.yml index 634b074c0..bbd461c89 100644 --- a/.github/workflows/__go-tracing-custom-build-steps.yml +++ b/.github/workflows/__go-tracing-custom-build-steps.yml @@ -91,7 +91,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__go-tracing-legacy-workflow.yml b/.github/workflows/__go-tracing-legacy-workflow.yml index 8168e3b10..feedfdff5 100644 --- a/.github/workflows/__go-tracing-legacy-workflow.yml +++ b/.github/workflows/__go-tracing-legacy-workflow.yml @@ -91,7 +91,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__init-with-registries.yml b/.github/workflows/__init-with-registries.yml index bbbc55bf1..8403d63e4 100644 --- a/.github/workflows/__init-with-registries.yml +++ b/.github/workflows/__init-with-registries.yml @@ -52,7 +52,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__javascript-source-root.yml b/.github/workflows/__javascript-source-root.yml index e6c883966..97caa3a69 100644 --- a/.github/workflows/__javascript-source-root.yml +++ b/.github/workflows/__javascript-source-root.yml @@ -51,7 +51,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__job-run-uuid-sarif.yml b/.github/workflows/__job-run-uuid-sarif.yml index b9f3eed91..73cb295fe 100644 --- a/.github/workflows/__job-run-uuid-sarif.yml +++ b/.github/workflows/__job-run-uuid-sarif.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__language-aliases.yml b/.github/workflows/__language-aliases.yml index 5f95caa13..b8976bb5d 100644 --- a/.github/workflows/__language-aliases.yml +++ b/.github/workflows/__language-aliases.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__local-bundle.yml b/.github/workflows/__local-bundle.yml index 3fc89f381..094f22ebc 100644 --- a/.github/workflows/__local-bundle.yml +++ b/.github/workflows/__local-bundle.yml @@ -77,7 +77,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__multi-language-autodetect.yml b/.github/workflows/__multi-language-autodetect.yml index 3704cdbf5..8e0b44dbf 100644 --- a/.github/workflows/__multi-language-autodetect.yml +++ b/.github/workflows/__multi-language-autodetect.yml @@ -111,7 +111,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__overlay-init-fallback.yml b/.github/workflows/__overlay-init-fallback.yml index d85e58aa1..b843b8aac 100644 --- a/.github/workflows/__overlay-init-fallback.yml +++ b/.github/workflows/__overlay-init-fallback.yml @@ -49,7 +49,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__packaging-codescanning-config-inputs-js.yml b/.github/workflows/__packaging-codescanning-config-inputs-js.yml index 53f280ab9..63875502f 100644 --- a/.github/workflows/__packaging-codescanning-config-inputs-js.yml +++ b/.github/workflows/__packaging-codescanning-config-inputs-js.yml @@ -81,7 +81,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Install Node.js uses: actions/setup-node@v6 with: diff --git a/.github/workflows/__packaging-config-inputs-js.yml b/.github/workflows/__packaging-config-inputs-js.yml index 2b483b41a..7a7972971 100644 --- a/.github/workflows/__packaging-config-inputs-js.yml +++ b/.github/workflows/__packaging-config-inputs-js.yml @@ -71,7 +71,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Install Node.js uses: actions/setup-node@v6 with: diff --git a/.github/workflows/__packaging-config-js.yml b/.github/workflows/__packaging-config-js.yml index d45ca3b36..00a6fc9da 100644 --- a/.github/workflows/__packaging-config-js.yml +++ b/.github/workflows/__packaging-config-js.yml @@ -71,7 +71,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Install Node.js uses: actions/setup-node@v6 with: diff --git a/.github/workflows/__packaging-inputs-js.yml b/.github/workflows/__packaging-inputs-js.yml index 41ca571b8..ec3ef3d5b 100644 --- a/.github/workflows/__packaging-inputs-js.yml +++ b/.github/workflows/__packaging-inputs-js.yml @@ -71,7 +71,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Install Node.js uses: actions/setup-node@v6 with: diff --git a/.github/workflows/__quality-queries.yml b/.github/workflows/__quality-queries.yml index 2a30bfceb..caef10d27 100644 --- a/.github/workflows/__quality-queries.yml +++ b/.github/workflows/__quality-queries.yml @@ -63,7 +63,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__remote-config.yml b/.github/workflows/__remote-config.yml index 20a308e74..f39b6f6f9 100644 --- a/.github/workflows/__remote-config.yml +++ b/.github/workflows/__remote-config.yml @@ -79,7 +79,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__resolve-environment-action.yml b/.github/workflows/__resolve-environment-action.yml index 2203f3316..01e242ebb 100644 --- a/.github/workflows/__resolve-environment-action.yml +++ b/.github/workflows/__resolve-environment-action.yml @@ -51,7 +51,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__rubocop-multi-language.yml b/.github/workflows/__rubocop-multi-language.yml index a5e457bb7..8340feced 100644 --- a/.github/workflows/__rubocop-multi-language.yml +++ b/.github/workflows/__rubocop-multi-language.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__ruby.yml b/.github/workflows/__ruby.yml index 769a11925..3050bf735 100644 --- a/.github/workflows/__ruby.yml +++ b/.github/workflows/__ruby.yml @@ -57,7 +57,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__rust.yml b/.github/workflows/__rust.yml index d788e5226..352ffdee7 100644 --- a/.github/workflows/__rust.yml +++ b/.github/workflows/__rust.yml @@ -55,7 +55,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__split-workflow.yml b/.github/workflows/__split-workflow.yml index 3ffb09928..c02385a72 100644 --- a/.github/workflows/__split-workflow.yml +++ b/.github/workflows/__split-workflow.yml @@ -77,7 +77,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__start-proxy.yml b/.github/workflows/__start-proxy.yml index 26f118460..40a2993ca 100644 --- a/.github/workflows/__start-proxy.yml +++ b/.github/workflows/__start-proxy.yml @@ -51,7 +51,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__submit-sarif-failure.yml b/.github/workflows/__submit-sarif-failure.yml index 7383b52a8..60c351020 100644 --- a/.github/workflows/__submit-sarif-failure.yml +++ b/.github/workflows/__submit-sarif-failure.yml @@ -52,7 +52,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -60,7 +60,7 @@ jobs: version: ${{ matrix.version }} use-all-platform-bundle: 'false' setup-kotlin: 'true' - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 - uses: ./init with: languages: javascript diff --git a/.github/workflows/__swift-autobuild.yml b/.github/workflows/__swift-autobuild.yml index 9d18d0c97..447f4ea9a 100644 --- a/.github/workflows/__swift-autobuild.yml +++ b/.github/workflows/__swift-autobuild.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__swift-custom-build.yml b/.github/workflows/__swift-custom-build.yml index a1c5a556f..49f655880 100644 --- a/.github/workflows/__swift-custom-build.yml +++ b/.github/workflows/__swift-custom-build.yml @@ -71,7 +71,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__unset-environment.yml b/.github/workflows/__unset-environment.yml index c1a62b110..2d52f3a80 100644 --- a/.github/workflows/__unset-environment.yml +++ b/.github/workflows/__unset-environment.yml @@ -79,7 +79,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__upload-ref-sha-input.yml b/.github/workflows/__upload-ref-sha-input.yml index 1c2c5975d..8de95e42a 100644 --- a/.github/workflows/__upload-ref-sha-input.yml +++ b/.github/workflows/__upload-ref-sha-input.yml @@ -77,7 +77,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__upload-sarif.yml b/.github/workflows/__upload-sarif.yml index 361c8228f..c4c31e4e4 100644 --- a/.github/workflows/__upload-sarif.yml +++ b/.github/workflows/__upload-sarif.yml @@ -84,7 +84,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__with-checkout-path.yml b/.github/workflows/__with-checkout-path.yml index 5aa2b631c..0ebb46432 100644 --- a/.github/workflows/__with-checkout-path.yml +++ b/.github/workflows/__with-checkout-path.yml @@ -77,7 +77,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -107,7 +107,7 @@ jobs: rm -rf ./* .github .git # Check out the actions repo again, but at a different location. # choose an arbitrary SHA so that we can later test that the commit_oid is not from main - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 with: ref: 474bbf07f9247ffe1856c6a0f94aeeb10e7afee6 path: x/y/z/some-path diff --git a/.github/workflows/check-expected-release-files.yml b/.github/workflows/check-expected-release-files.yml index a066cbde5..c0dd21af6 100644 --- a/.github/workflows/check-expected-release-files.yml +++ b/.github/workflows/check-expected-release-files.yml @@ -22,7 +22,7 @@ jobs: steps: - name: Checkout CodeQL Action - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Check Expected Release Files run: | bundle_version="$(cat "./src/defaults.json" | jq -r ".bundleVersion")" diff --git a/.github/workflows/codeql.yml b/.github/workflows/codeql.yml index 24dace33c..8ea440089 100644 --- a/.github/workflows/codeql.yml +++ b/.github/workflows/codeql.yml @@ -32,7 +32,7 @@ jobs: contents: read steps: - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 - name: Init with default CodeQL bundle from the VM image id: init-default uses: ./init @@ -91,7 +91,7 @@ jobs: steps: - name: Checkout - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Initialize CodeQL uses: ./init id: init @@ -128,7 +128,7 @@ jobs: steps: - name: Checkout - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Initialize CodeQL uses: ./init with: diff --git a/.github/workflows/codescanning-config-cli.yml b/.github/workflows/codescanning-config-cli.yml index 5ae95f68a..3c97239d5 100644 --- a/.github/workflows/codescanning-config-cli.yml +++ b/.github/workflows/codescanning-config-cli.yml @@ -53,7 +53,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Set up Node.js uses: actions/setup-node@v6 diff --git a/.github/workflows/debug-artifacts-failure-safe.yml b/.github/workflows/debug-artifacts-failure-safe.yml index 768f88f96..3f710863e 100644 --- a/.github/workflows/debug-artifacts-failure-safe.yml +++ b/.github/workflows/debug-artifacts-failure-safe.yml @@ -45,7 +45,7 @@ jobs: - name: Dump GitHub event run: cat "${GITHUB_EVENT_PATH}" - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/debug-artifacts-safe.yml b/.github/workflows/debug-artifacts-safe.yml index e33d70cc3..7cee73cbe 100644 --- a/.github/workflows/debug-artifacts-safe.yml +++ b/.github/workflows/debug-artifacts-safe.yml @@ -41,7 +41,7 @@ jobs: runs-on: ubuntu-latest steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/post-release-mergeback.yml b/.github/workflows/post-release-mergeback.yml index 1731a78ff..71dae6a3a 100644 --- a/.github/workflows/post-release-mergeback.yml +++ b/.github/workflows/post-release-mergeback.yml @@ -44,7 +44,7 @@ jobs: GITHUB_CONTEXT: '${{ toJson(github) }}' run: echo "${GITHUB_CONTEXT}" - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 with: fetch-depth: 0 # ensure we have all tags and can push commits - uses: actions/setup-node@v6 diff --git a/.github/workflows/pr-checks.yml b/.github/workflows/pr-checks.yml index 9aa0355c1..5badaab81 100644 --- a/.github/workflows/pr-checks.yml +++ b/.github/workflows/pr-checks.yml @@ -32,7 +32,7 @@ jobs: if: runner.os == 'Windows' run: git config --global core.autocrlf false - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 - name: Set up Node.js uses: actions/setup-node@v6 @@ -91,7 +91,7 @@ jobs: contents: read steps: - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 - id: head-version name: Verify all Actions use the same Node version run: | @@ -106,7 +106,7 @@ jobs: - id: checkout-base name: 'Backport: Check out base ref' if: ${{ startsWith(github.head_ref, 'backport-') }} - uses: actions/checkout@v5 + uses: actions/checkout@v6 with: ref: ${{ env.BASE_REF }} diff --git a/.github/workflows/prepare-release.yml b/.github/workflows/prepare-release.yml index dad6fce39..7e9486bb4 100644 --- a/.github/workflows/prepare-release.yml +++ b/.github/workflows/prepare-release.yml @@ -44,7 +44,7 @@ jobs: steps: - name: Checkout repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 with: fetch-depth: 0 # Need full history for calculation of diffs diff --git a/.github/workflows/publish-immutable-action.yml b/.github/workflows/publish-immutable-action.yml index c6084573c..e14bc30bc 100644 --- a/.github/workflows/publish-immutable-action.yml +++ b/.github/workflows/publish-immutable-action.yml @@ -20,7 +20,7 @@ jobs: steps: - name: Checkout repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Publish immutable release id: publish diff --git a/.github/workflows/python312-windows.yml b/.github/workflows/python312-windows.yml index aa2a03420..8ef1be866 100644 --- a/.github/workflows/python312-windows.yml +++ b/.github/workflows/python312-windows.yml @@ -31,7 +31,7 @@ jobs: with: python-version: 3.12 - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 - name: Prepare test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/query-filters.yml b/.github/workflows/query-filters.yml index 3e17d989e..90e702c93 100644 --- a/.github/workflows/query-filters.yml +++ b/.github/workflows/query-filters.yml @@ -29,7 +29,7 @@ jobs: contents: read # This permission is needed to allow the GitHub Actions workflow to read the contents of the repository. steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Install Node.js uses: actions/setup-node@v6 diff --git a/.github/workflows/rebuild.yml b/.github/workflows/rebuild.yml index e7b9022be..9740a0d16 100644 --- a/.github/workflows/rebuild.yml +++ b/.github/workflows/rebuild.yml @@ -24,7 +24,7 @@ jobs: pull-requests: write # needed to comment on the PR steps: - name: Checkout - uses: actions/checkout@v5 + uses: actions/checkout@v6 with: fetch-depth: 0 ref: ${{ env.HEAD_REF }} diff --git a/.github/workflows/rollback-release.yml b/.github/workflows/rollback-release.yml index 8d8e872fa..4f419b82b 100644 --- a/.github/workflows/rollback-release.yml +++ b/.github/workflows/rollback-release.yml @@ -52,7 +52,7 @@ jobs: steps: - name: Checkout repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 with: fetch-depth: 0 # Need full history for calculation of diffs diff --git a/.github/workflows/test-codeql-bundle-all.yml b/.github/workflows/test-codeql-bundle-all.yml index 6465d6a1d..395288275 100644 --- a/.github/workflows/test-codeql-bundle-all.yml +++ b/.github/workflows/test-codeql-bundle-all.yml @@ -36,7 +36,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/update-bundle.yml b/.github/workflows/update-bundle.yml index 184c339ff..951b89066 100644 --- a/.github/workflows/update-bundle.yml +++ b/.github/workflows/update-bundle.yml @@ -33,7 +33,7 @@ jobs: GITHUB_CONTEXT: '${{ toJson(github) }}' run: echo "$GITHUB_CONTEXT" - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 - name: Update git config run: | diff --git a/.github/workflows/update-release-branch.yml b/.github/workflows/update-release-branch.yml index 830ed7c2a..bd678c655 100644 --- a/.github/workflows/update-release-branch.yml +++ b/.github/workflows/update-release-branch.yml @@ -38,7 +38,7 @@ jobs: contents: write # needed to push commits pull-requests: write # needed to create pull request steps: - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 with: fetch-depth: 0 # Need full history for calculation of diffs - uses: ./.github/actions/release-initialise @@ -100,7 +100,7 @@ jobs: private-key: ${{ secrets.AUTOMATION_PRIVATE_KEY }} - name: Checkout - uses: actions/checkout@v5 + uses: actions/checkout@v6 with: fetch-depth: 0 # Need full history for calculation of diffs token: ${{ steps.app-token.outputs.token }} diff --git a/.github/workflows/update-supported-enterprise-server-versions.yml b/.github/workflows/update-supported-enterprise-server-versions.yml index 421a63c69..4cead58f4 100644 --- a/.github/workflows/update-supported-enterprise-server-versions.yml +++ b/.github/workflows/update-supported-enterprise-server-versions.yml @@ -27,9 +27,9 @@ jobs: with: python-version: "3.13" - name: Checkout CodeQL Action - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Checkout Enterprise Releases - uses: actions/checkout@v5 + uses: actions/checkout@v6 with: repository: github/enterprise-releases token: ${{ secrets.ENTERPRISE_RELEASE_TOKEN }} From 8484f54a0a681dd8cf94876c4283a3e8ea0e6178 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Mon, 24 Nov 2025 18:02:41 +0000 Subject: [PATCH 51/51] Rebuild --- pr-checks/checks/submit-sarif-failure.yml | 2 +- pr-checks/checks/with-checkout-path.yml | 2 +- pr-checks/sync.py | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/pr-checks/checks/submit-sarif-failure.yml b/pr-checks/checks/submit-sarif-failure.yml index 97332e4c9..5db63bb81 100644 --- a/pr-checks/checks/submit-sarif-failure.yml +++ b/pr-checks/checks/submit-sarif-failure.yml @@ -18,7 +18,7 @@ permissions: security-events: write # needed to upload the SARIF file steps: - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 - uses: ./init with: languages: javascript diff --git a/pr-checks/checks/with-checkout-path.yml b/pr-checks/checks/with-checkout-path.yml index 5cdd02c0d..230e342e3 100644 --- a/pr-checks/checks/with-checkout-path.yml +++ b/pr-checks/checks/with-checkout-path.yml @@ -14,7 +14,7 @@ steps: rm -rf ./* .github .git # Check out the actions repo again, but at a different location. # choose an arbitrary SHA so that we can later test that the commit_oid is not from main - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 with: ref: 474bbf07f9247ffe1856c6a0f94aeeb10e7afee6 path: x/y/z/some-path diff --git a/pr-checks/sync.py b/pr-checks/sync.py index 77816be76..7d412e9b0 100755 --- a/pr-checks/sync.py +++ b/pr-checks/sync.py @@ -107,7 +107,7 @@ for file in sorted((this_dir / 'checks').glob('*.yml')): steps = [ { 'name': 'Check out repository', - 'uses': 'actions/checkout@v5' + 'uses': 'actions/checkout@v6' }, ]