diff --git a/.github/pull_request_template.md b/.github/pull_request_template.md index 3b632d8f3..3c6d14f71 100644 --- a/.github/pull_request_template.md +++ b/.github/pull_request_template.md @@ -18,14 +18,25 @@ For internal use only. Please select the risk level of this change: #### Which use cases does this change impact? - + -- **Advanced setup** - Impacts users who have custom workflows. -- **Default setup** - Impacts users who use default setup. -- **Code Scanning** - Impacts Code Scanning (i.e. `analysis-kinds: code-scanning`). -- **Code Quality** - Impacts Code Quality (i.e. `analysis-kinds: code-quality`). -- **Third-party analyses** - Impacts third-party analyses (i.e. `upload-sarif`). -- **GHES** - Impacts GitHub Enterprise Server. +Workflow types: + +- **Advanced setup** - Impacts users who have custom CodeQL workflows. +- **Managed** - Impacts users with `dynamic` workflows (Default Setup, CCR, ...). + +Products: + +- **Code Scanning** - The changes impact analyses when `analysis-kinds: code-scanning`. +- **Code Quality** - The changes impact analyses when `analysis-kinds: code-quality`. +- **CCR** - The changes impact analyses for Copilot Code Reviews. +- **Third-party analyses** - The changes affect the `upload-sarif` action. + +Environments: + +- **Dotcom** - Impacts CodeQL workflows on `github.com`. +- **GHES** - Impacts CodeQL workflows on GitHub Enterprise Server. +- **Testing/None** - This change does not impact any CodeQL workflows in production. #### How did/will you validate this change? @@ -54,6 +65,15 @@ For internal use only. Please select the risk level of this change: - **Alerts** - New or existing monitors will trip if something goes wrong with this change. - **Other** - Please provide details. +#### Are there any special considerations for merging or releasing this change? + + + +- **No special considerations** - This change can be merged at any time. +- **Special considerations** - This change should only be merged once certain preconditions are met. Please provide details of those or link to this PR from an internal issue. + ### Merge / deployment checklist - Confirm this change is backwards compatible with existing workflows. diff --git a/.github/workflows/__all-platform-bundle.yml b/.github/workflows/__all-platform-bundle.yml index 89138c523..2340be49c 100644 --- a/.github/workflows/__all-platform-bundle.yml +++ b/.github/workflows/__all-platform-bundle.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -61,7 +71,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -74,6 +84,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - id: init uses: ./../action/init with: diff --git a/.github/workflows/__analyze-ref-input.yml b/.github/workflows/__analyze-ref-input.yml index 9cf85b0c5..161942723 100644 --- a/.github/workflows/__analyze-ref-input.yml +++ b/.github/workflows/__analyze-ref-input.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -67,7 +77,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -85,6 +95,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: tools: ${{ steps.prepare-test.outputs.tools-url }} diff --git a/.github/workflows/__autobuild-action.yml b/.github/workflows/__autobuild-action.yml index c31576339..08470bcff 100644 --- a/.github/workflows/__autobuild-action.yml +++ b/.github/workflows/__autobuild-action.yml @@ -21,9 +21,19 @@ on: schedule: - cron: '0 5 * * *' workflow_dispatch: - inputs: {} + inputs: + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: - inputs: {} + inputs: + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -51,7 +61,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -59,6 +69,10 @@ jobs: version: ${{ matrix.version }} use-all-platform-bundle: 'false' setup-kotlin: 'true' + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: languages: csharp diff --git a/.github/workflows/__autobuild-direct-tracing-with-working-dir.yml b/.github/workflows/__autobuild-direct-tracing-with-working-dir.yml index c1de5c19d..9607fce18 100644 --- a/.github/workflows/__autobuild-direct-tracing-with-working-dir.yml +++ b/.github/workflows/__autobuild-direct-tracing-with-working-dir.yml @@ -63,7 +63,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__autobuild-working-dir.yml b/.github/workflows/__autobuild-working-dir.yml index 3a3ca9e5f..e9d1d7d5d 100644 --- a/.github/workflows/__autobuild-working-dir.yml +++ b/.github/workflows/__autobuild-working-dir.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__build-mode-autobuild.yml b/.github/workflows/__build-mode-autobuild.yml index 878c941a4..87ed95e1e 100644 --- a/.github/workflows/__build-mode-autobuild.yml +++ b/.github/workflows/__build-mode-autobuild.yml @@ -63,7 +63,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__build-mode-manual.yml b/.github/workflows/__build-mode-manual.yml index e0dc25f88..c164a1a7b 100644 --- a/.github/workflows/__build-mode-manual.yml +++ b/.github/workflows/__build-mode-manual.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -57,7 +67,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -70,6 +80,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init id: init with: diff --git a/.github/workflows/__build-mode-none.yml b/.github/workflows/__build-mode-none.yml index 7584f9065..7bb121810 100644 --- a/.github/workflows/__build-mode-none.yml +++ b/.github/workflows/__build-mode-none.yml @@ -49,7 +49,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__build-mode-rollback.yml b/.github/workflows/__build-mode-rollback.yml index c1f3ccd0c..e9d85968c 100644 --- a/.github/workflows/__build-mode-rollback.yml +++ b/.github/workflows/__build-mode-rollback.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__bundle-from-toolcache.yml b/.github/workflows/__bundle-from-toolcache.yml index 639595af5..96858acd1 100644 --- a/.github/workflows/__bundle-from-toolcache.yml +++ b/.github/workflows/__bundle-from-toolcache.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__bundle-toolcache.yml b/.github/workflows/__bundle-toolcache.yml index de3826b65..59d06b49b 100644 --- a/.github/workflows/__bundle-toolcache.yml +++ b/.github/workflows/__bundle-toolcache.yml @@ -51,7 +51,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__bundle-zstd.yml b/.github/workflows/__bundle-zstd.yml index f5b1ab3aa..18185ada3 100644 --- a/.github/workflows/__bundle-zstd.yml +++ b/.github/workflows/__bundle-zstd.yml @@ -51,7 +51,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__cleanup-db-cluster-dir.yml b/.github/workflows/__cleanup-db-cluster-dir.yml index dfe53c67c..8bf4659ae 100644 --- a/.github/workflows/__cleanup-db-cluster-dir.yml +++ b/.github/workflows/__cleanup-db-cluster-dir.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__config-export.yml b/.github/workflows/__config-export.yml index f01c4ae3d..1c9895854 100644 --- a/.github/workflows/__config-export.yml +++ b/.github/workflows/__config-export.yml @@ -49,7 +49,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__config-input.yml b/.github/workflows/__config-input.yml index 59db10d4d..2a006be21 100644 --- a/.github/workflows/__config-input.yml +++ b/.github/workflows/__config-input.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Install Node.js uses: actions/setup-node@v6 with: diff --git a/.github/workflows/__cpp-deptrace-disabled.yml b/.github/workflows/__cpp-deptrace-disabled.yml index 122159236..2116e5c4f 100644 --- a/.github/workflows/__cpp-deptrace-disabled.yml +++ b/.github/workflows/__cpp-deptrace-disabled.yml @@ -51,7 +51,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__cpp-deptrace-enabled-on-macos.yml b/.github/workflows/__cpp-deptrace-enabled-on-macos.yml index b9669b870..1039cc321 100644 --- a/.github/workflows/__cpp-deptrace-enabled-on-macos.yml +++ b/.github/workflows/__cpp-deptrace-enabled-on-macos.yml @@ -49,7 +49,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__cpp-deptrace-enabled.yml b/.github/workflows/__cpp-deptrace-enabled.yml index bf155a64d..9a57d0041 100644 --- a/.github/workflows/__cpp-deptrace-enabled.yml +++ b/.github/workflows/__cpp-deptrace-enabled.yml @@ -51,7 +51,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__diagnostics-export.yml b/.github/workflows/__diagnostics-export.yml index 9251e04a8..8c05b6d92 100644 --- a/.github/workflows/__diagnostics-export.yml +++ b/.github/workflows/__diagnostics-export.yml @@ -49,7 +49,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__export-file-baseline-information.yml b/.github/workflows/__export-file-baseline-information.yml index 66274f26b..7ebf51f3f 100644 --- a/.github/workflows/__export-file-baseline-information.yml +++ b/.github/workflows/__export-file-baseline-information.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -61,7 +71,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -74,6 +84,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init id: init with: diff --git a/.github/workflows/__extractor-ram-threads.yml b/.github/workflows/__extractor-ram-threads.yml index 2d8316f52..09c1cbbf4 100644 --- a/.github/workflows/__extractor-ram-threads.yml +++ b/.github/workflows/__extractor-ram-threads.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__global-proxy.yml b/.github/workflows/__global-proxy.yml index bd5d64b5f..35f1f08fc 100644 --- a/.github/workflows/__global-proxy.yml +++ b/.github/workflows/__global-proxy.yml @@ -61,7 +61,7 @@ jobs: apt install -y gh env: {} - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__go-custom-queries.yml b/.github/workflows/__go-custom-queries.yml index 1b5b7b915..32912ee07 100644 --- a/.github/workflows/__go-custom-queries.yml +++ b/.github/workflows/__go-custom-queries.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -59,7 +69,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -72,6 +82,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: languages: go diff --git a/.github/workflows/__go-indirect-tracing-workaround-diagnostic.yml b/.github/workflows/__go-indirect-tracing-workaround-diagnostic.yml index 061ad4254..b140b13c4 100644 --- a/.github/workflows/__go-indirect-tracing-workaround-diagnostic.yml +++ b/.github/workflows/__go-indirect-tracing-workaround-diagnostic.yml @@ -57,7 +57,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__go-indirect-tracing-workaround-no-file-program.yml b/.github/workflows/__go-indirect-tracing-workaround-no-file-program.yml index 0a347c65c..d6cf269d7 100644 --- a/.github/workflows/__go-indirect-tracing-workaround-no-file-program.yml +++ b/.github/workflows/__go-indirect-tracing-workaround-no-file-program.yml @@ -57,7 +57,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__go-indirect-tracing-workaround.yml b/.github/workflows/__go-indirect-tracing-workaround.yml index bb811d4d5..8b0c3d13d 100644 --- a/.github/workflows/__go-indirect-tracing-workaround.yml +++ b/.github/workflows/__go-indirect-tracing-workaround.yml @@ -57,7 +57,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__go-tracing-autobuilder.yml b/.github/workflows/__go-tracing-autobuilder.yml index 6d4cc91cc..0d2db3734 100644 --- a/.github/workflows/__go-tracing-autobuilder.yml +++ b/.github/workflows/__go-tracing-autobuilder.yml @@ -91,7 +91,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__go-tracing-custom-build-steps.yml b/.github/workflows/__go-tracing-custom-build-steps.yml index 634b074c0..bbd461c89 100644 --- a/.github/workflows/__go-tracing-custom-build-steps.yml +++ b/.github/workflows/__go-tracing-custom-build-steps.yml @@ -91,7 +91,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__go-tracing-legacy-workflow.yml b/.github/workflows/__go-tracing-legacy-workflow.yml index 8168e3b10..feedfdff5 100644 --- a/.github/workflows/__go-tracing-legacy-workflow.yml +++ b/.github/workflows/__go-tracing-legacy-workflow.yml @@ -91,7 +91,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__go.yml b/.github/workflows/__go.yml index e8bf8837e..fb27da710 100644 --- a/.github/workflows/__go.yml +++ b/.github/workflows/__go.yml @@ -18,6 +18,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x jobs: go-custom-queries: name: 'Go: Custom queries' @@ -27,6 +32,7 @@ jobs: uses: ./.github/workflows/__go-custom-queries.yml with: go-version: ${{ inputs.go-version }} + dotnet-version: ${{ inputs.dotnet-version }} go-indirect-tracing-workaround-diagnostic: name: 'Go: diagnostic when Go is changed after init step' permissions: diff --git a/.github/workflows/__init-with-registries.yml b/.github/workflows/__init-with-registries.yml index bbbc55bf1..8403d63e4 100644 --- a/.github/workflows/__init-with-registries.yml +++ b/.github/workflows/__init-with-registries.yml @@ -52,7 +52,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__javascript-source-root.yml b/.github/workflows/__javascript-source-root.yml index e6c883966..97caa3a69 100644 --- a/.github/workflows/__javascript-source-root.yml +++ b/.github/workflows/__javascript-source-root.yml @@ -51,7 +51,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__job-run-uuid-sarif.yml b/.github/workflows/__job-run-uuid-sarif.yml index b9f3eed91..73cb295fe 100644 --- a/.github/workflows/__job-run-uuid-sarif.yml +++ b/.github/workflows/__job-run-uuid-sarif.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__language-aliases.yml b/.github/workflows/__language-aliases.yml index 5f95caa13..b8976bb5d 100644 --- a/.github/workflows/__language-aliases.yml +++ b/.github/workflows/__language-aliases.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__local-bundle.yml b/.github/workflows/__local-bundle.yml index 159d8c455..094f22ebc 100644 --- a/.github/workflows/__local-bundle.yml +++ b/.github/workflows/__local-bundle.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -67,7 +77,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -85,6 +95,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - name: Fetch latest CodeQL bundle run: | wget https://github.com/github/codeql-action/releases/latest/download/codeql-bundle-linux64.tar.zst diff --git a/.github/workflows/__multi-language-autodetect.yml b/.github/workflows/__multi-language-autodetect.yml index e9272f893..8e0b44dbf 100644 --- a/.github/workflows/__multi-language-autodetect.yml +++ b/.github/workflows/__multi-language-autodetect.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -101,7 +111,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -119,6 +129,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - name: Use Xcode 16 if: runner.os == 'macOS' && matrix.version != 'nightly-latest' run: sudo xcode-select -s "/Applications/Xcode_16.app" diff --git a/.github/workflows/__overlay-init-fallback.yml b/.github/workflows/__overlay-init-fallback.yml index d85e58aa1..b843b8aac 100644 --- a/.github/workflows/__overlay-init-fallback.yml +++ b/.github/workflows/__overlay-init-fallback.yml @@ -49,7 +49,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__packaging-codescanning-config-inputs-js.yml b/.github/workflows/__packaging-codescanning-config-inputs-js.yml index a5038cefc..63875502f 100644 --- a/.github/workflows/__packaging-codescanning-config-inputs-js.yml +++ b/.github/workflows/__packaging-codescanning-config-inputs-js.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -71,7 +81,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Install Node.js uses: actions/setup-node@v6 with: @@ -96,6 +106,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: config-file: .github/codeql/codeql-config-packaging3.yml diff --git a/.github/workflows/__packaging-config-inputs-js.yml b/.github/workflows/__packaging-config-inputs-js.yml index 9c8483c74..7a7972971 100644 --- a/.github/workflows/__packaging-config-inputs-js.yml +++ b/.github/workflows/__packaging-config-inputs-js.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -61,7 +71,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Install Node.js uses: actions/setup-node@v6 with: @@ -81,6 +91,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: config-file: .github/codeql/codeql-config-packaging3.yml diff --git a/.github/workflows/__packaging-config-js.yml b/.github/workflows/__packaging-config-js.yml index b7608793a..00a6fc9da 100644 --- a/.github/workflows/__packaging-config-js.yml +++ b/.github/workflows/__packaging-config-js.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -61,7 +71,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Install Node.js uses: actions/setup-node@v6 with: @@ -81,6 +91,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: config-file: .github/codeql/codeql-config-packaging.yml diff --git a/.github/workflows/__packaging-inputs-js.yml b/.github/workflows/__packaging-inputs-js.yml index 0a814a050..ec3ef3d5b 100644 --- a/.github/workflows/__packaging-inputs-js.yml +++ b/.github/workflows/__packaging-inputs-js.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -61,7 +71,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Install Node.js uses: actions/setup-node@v6 with: @@ -81,6 +91,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: config-file: .github/codeql/codeql-config-packaging2.yml diff --git a/.github/workflows/__quality-queries.yml b/.github/workflows/__quality-queries.yml index 2a30bfceb..caef10d27 100644 --- a/.github/workflows/__quality-queries.yml +++ b/.github/workflows/__quality-queries.yml @@ -63,7 +63,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__remote-config.yml b/.github/workflows/__remote-config.yml index b59da417e..f39b6f6f9 100644 --- a/.github/workflows/__remote-config.yml +++ b/.github/workflows/__remote-config.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -69,7 +79,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -87,6 +97,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: tools: ${{ steps.prepare-test.outputs.tools-url }} diff --git a/.github/workflows/__resolve-environment-action.yml b/.github/workflows/__resolve-environment-action.yml index 2203f3316..01e242ebb 100644 --- a/.github/workflows/__resolve-environment-action.yml +++ b/.github/workflows/__resolve-environment-action.yml @@ -51,7 +51,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__rubocop-multi-language.yml b/.github/workflows/__rubocop-multi-language.yml index 5442459ff..8340feced 100644 --- a/.github/workflows/__rubocop-multi-language.yml +++ b/.github/workflows/__rubocop-multi-language.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -56,7 +56,7 @@ jobs: use-all-platform-bundle: 'false' setup-kotlin: 'true' - name: Set up Ruby - uses: ruby/setup-ruby@d5126b9b3579e429dd52e51e68624dda2e05be25 # v1.267.0 + uses: ruby/setup-ruby@8aeb6ff8030dd539317f8e1769a044873b56ea71 # v1.268.0 with: ruby-version: 2.6 - name: Install Code Scanning integration diff --git a/.github/workflows/__ruby.yml b/.github/workflows/__ruby.yml index 769a11925..3050bf735 100644 --- a/.github/workflows/__ruby.yml +++ b/.github/workflows/__ruby.yml @@ -57,7 +57,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__rust.yml b/.github/workflows/__rust.yml index d788e5226..352ffdee7 100644 --- a/.github/workflows/__rust.yml +++ b/.github/workflows/__rust.yml @@ -55,7 +55,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__split-workflow.yml b/.github/workflows/__split-workflow.yml index e916b36cc..c02385a72 100644 --- a/.github/workflows/__split-workflow.yml +++ b/.github/workflows/__split-workflow.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -67,7 +77,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -80,6 +90,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: config-file: .github/codeql/codeql-config-packaging3.yml diff --git a/.github/workflows/__start-proxy.yml b/.github/workflows/__start-proxy.yml index 26f118460..40a2993ca 100644 --- a/.github/workflows/__start-proxy.yml +++ b/.github/workflows/__start-proxy.yml @@ -51,7 +51,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__submit-sarif-failure.yml b/.github/workflows/__submit-sarif-failure.yml index 7383b52a8..60c351020 100644 --- a/.github/workflows/__submit-sarif-failure.yml +++ b/.github/workflows/__submit-sarif-failure.yml @@ -52,7 +52,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -60,7 +60,7 @@ jobs: version: ${{ matrix.version }} use-all-platform-bundle: 'false' setup-kotlin: 'true' - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 - uses: ./init with: languages: javascript diff --git a/.github/workflows/__swift-autobuild.yml b/.github/workflows/__swift-autobuild.yml index 9d18d0c97..447f4ea9a 100644 --- a/.github/workflows/__swift-autobuild.yml +++ b/.github/workflows/__swift-autobuild.yml @@ -47,7 +47,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/__swift-custom-build.yml b/.github/workflows/__swift-custom-build.yml index 32ce33a7f..49f655880 100644 --- a/.github/workflows/__swift-custom-build.yml +++ b/.github/workflows/__swift-custom-build.yml @@ -27,6 +27,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -34,6 +39,11 @@ on: description: The version of Go to install required: false default: '>=1.21.0' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -61,7 +71,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -74,6 +84,10 @@ jobs: with: go-version: ${{ inputs.go-version || '>=1.21.0' }} cache: false + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - name: Use Xcode 16 if: runner.os == 'macOS' && matrix.version != 'nightly-latest' run: sudo xcode-select -s "/Applications/Xcode_16.app" diff --git a/.github/workflows/__unset-environment.yml b/.github/workflows/__unset-environment.yml index ced3b33fc..2d52f3a80 100644 --- a/.github/workflows/__unset-environment.yml +++ b/.github/workflows/__unset-environment.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -69,7 +79,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -87,6 +97,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init id: init with: diff --git a/.github/workflows/__upload-ref-sha-input.yml b/.github/workflows/__upload-ref-sha-input.yml index 8f1fd4917..8de95e42a 100644 --- a/.github/workflows/__upload-ref-sha-input.yml +++ b/.github/workflows/__upload-ref-sha-input.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -67,7 +77,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -85,6 +95,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: tools: ${{ steps.prepare-test.outputs.tools-url }} diff --git a/.github/workflows/__upload-sarif.yml b/.github/workflows/__upload-sarif.yml index 77cd057f9..c4c31e4e4 100644 --- a/.github/workflows/__upload-sarif.yml +++ b/.github/workflows/__upload-sarif.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -74,7 +84,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -92,6 +102,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - uses: ./../action/init with: tools: ${{ steps.prepare-test.outputs.tools-url }} diff --git a/.github/workflows/__with-checkout-path.yml b/.github/workflows/__with-checkout-path.yml index 1b4002ffb..0ebb46432 100644 --- a/.github/workflows/__with-checkout-path.yml +++ b/.github/workflows/__with-checkout-path.yml @@ -32,6 +32,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x workflow_call: inputs: go-version: @@ -44,6 +49,11 @@ on: description: The version of Python to install required: false default: '3.13' + dotnet-version: + type: string + description: The version of .NET to install + required: false + default: 9.x defaults: run: shell: bash @@ -67,7 +77,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -85,6 +95,10 @@ jobs: uses: actions/setup-python@v6 with: python-version: ${{ inputs.python-version || '3.13' }} + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: ${{ inputs.dotnet-version || '9.x' }} - name: Delete original checkout run: | # delete the original checkout so we don't accidentally use it. @@ -93,7 +107,7 @@ jobs: rm -rf ./* .github .git # Check out the actions repo again, but at a different location. # choose an arbitrary SHA so that we can later test that the commit_oid is not from main - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 with: ref: 474bbf07f9247ffe1856c6a0f94aeeb10e7afee6 path: x/y/z/some-path diff --git a/.github/workflows/check-expected-release-files.yml b/.github/workflows/check-expected-release-files.yml index a066cbde5..c0dd21af6 100644 --- a/.github/workflows/check-expected-release-files.yml +++ b/.github/workflows/check-expected-release-files.yml @@ -22,7 +22,7 @@ jobs: steps: - name: Checkout CodeQL Action - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Check Expected Release Files run: | bundle_version="$(cat "./src/defaults.json" | jq -r ".bundleVersion")" diff --git a/.github/workflows/codeql.yml b/.github/workflows/codeql.yml index 24dace33c..8ea440089 100644 --- a/.github/workflows/codeql.yml +++ b/.github/workflows/codeql.yml @@ -32,7 +32,7 @@ jobs: contents: read steps: - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 - name: Init with default CodeQL bundle from the VM image id: init-default uses: ./init @@ -91,7 +91,7 @@ jobs: steps: - name: Checkout - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Initialize CodeQL uses: ./init id: init @@ -128,7 +128,7 @@ jobs: steps: - name: Checkout - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Initialize CodeQL uses: ./init with: diff --git a/.github/workflows/codescanning-config-cli.yml b/.github/workflows/codescanning-config-cli.yml index 5ae95f68a..3c97239d5 100644 --- a/.github/workflows/codescanning-config-cli.yml +++ b/.github/workflows/codescanning-config-cli.yml @@ -53,7 +53,7 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Set up Node.js uses: actions/setup-node@v6 diff --git a/.github/workflows/debug-artifacts-failure-safe.yml b/.github/workflows/debug-artifacts-failure-safe.yml index 1a09b3d9e..3f710863e 100644 --- a/.github/workflows/debug-artifacts-failure-safe.yml +++ b/.github/workflows/debug-artifacts-failure-safe.yml @@ -45,7 +45,7 @@ jobs: - name: Dump GitHub event run: cat "${GITHUB_EVENT_PATH}" - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -54,6 +54,10 @@ jobs: - uses: actions/setup-go@v6 with: go-version: ^1.13.1 + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: '9.x' - uses: ./../action/init with: tools: ${{ steps.prepare-test.outputs.tools-url }} diff --git a/.github/workflows/debug-artifacts-safe.yml b/.github/workflows/debug-artifacts-safe.yml index ea513521f..7cee73cbe 100644 --- a/.github/workflows/debug-artifacts-safe.yml +++ b/.github/workflows/debug-artifacts-safe.yml @@ -41,7 +41,7 @@ jobs: runs-on: ubuntu-latest steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test @@ -50,6 +50,10 @@ jobs: - uses: actions/setup-go@v6 with: go-version: ^1.13.1 + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: '9.x' - uses: ./../action/init id: init with: diff --git a/.github/workflows/post-release-mergeback.yml b/.github/workflows/post-release-mergeback.yml index 1731a78ff..601e3d1f8 100644 --- a/.github/workflows/post-release-mergeback.yml +++ b/.github/workflows/post-release-mergeback.yml @@ -44,7 +44,7 @@ jobs: GITHUB_CONTEXT: '${{ toJson(github) }}' run: echo "${GITHUB_CONTEXT}" - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 with: fetch-depth: 0 # ensure we have all tags and can push commits - uses: actions/setup-node@v6 @@ -142,7 +142,7 @@ jobs: token: "${{ secrets.GITHUB_TOKEN }}" - name: Generate token - uses: actions/create-github-app-token@v2.1.4 + uses: actions/create-github-app-token@v2.2.0 id: app-token with: app-id: ${{ vars.AUTOMATION_APP_ID }} diff --git a/.github/workflows/pr-checks.yml b/.github/workflows/pr-checks.yml index 9aa0355c1..5badaab81 100644 --- a/.github/workflows/pr-checks.yml +++ b/.github/workflows/pr-checks.yml @@ -32,7 +32,7 @@ jobs: if: runner.os == 'Windows' run: git config --global core.autocrlf false - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 - name: Set up Node.js uses: actions/setup-node@v6 @@ -91,7 +91,7 @@ jobs: contents: read steps: - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 - id: head-version name: Verify all Actions use the same Node version run: | @@ -106,7 +106,7 @@ jobs: - id: checkout-base name: 'Backport: Check out base ref' if: ${{ startsWith(github.head_ref, 'backport-') }} - uses: actions/checkout@v5 + uses: actions/checkout@v6 with: ref: ${{ env.BASE_REF }} diff --git a/.github/workflows/prepare-release.yml b/.github/workflows/prepare-release.yml index dad6fce39..7e9486bb4 100644 --- a/.github/workflows/prepare-release.yml +++ b/.github/workflows/prepare-release.yml @@ -44,7 +44,7 @@ jobs: steps: - name: Checkout repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 with: fetch-depth: 0 # Need full history for calculation of diffs diff --git a/.github/workflows/publish-immutable-action.yml b/.github/workflows/publish-immutable-action.yml index c6084573c..e14bc30bc 100644 --- a/.github/workflows/publish-immutable-action.yml +++ b/.github/workflows/publish-immutable-action.yml @@ -20,7 +20,7 @@ jobs: steps: - name: Checkout repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Publish immutable release id: publish diff --git a/.github/workflows/python312-windows.yml b/.github/workflows/python312-windows.yml index aa2a03420..8ef1be866 100644 --- a/.github/workflows/python312-windows.yml +++ b/.github/workflows/python312-windows.yml @@ -31,7 +31,7 @@ jobs: with: python-version: 3.12 - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 - name: Prepare test uses: ./.github/actions/prepare-test diff --git a/.github/workflows/query-filters.yml b/.github/workflows/query-filters.yml index 3e17d989e..90e702c93 100644 --- a/.github/workflows/query-filters.yml +++ b/.github/workflows/query-filters.yml @@ -29,7 +29,7 @@ jobs: contents: read # This permission is needed to allow the GitHub Actions workflow to read the contents of the repository. steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Install Node.js uses: actions/setup-node@v6 diff --git a/.github/workflows/rebuild.yml b/.github/workflows/rebuild.yml index e7b9022be..9740a0d16 100644 --- a/.github/workflows/rebuild.yml +++ b/.github/workflows/rebuild.yml @@ -24,7 +24,7 @@ jobs: pull-requests: write # needed to comment on the PR steps: - name: Checkout - uses: actions/checkout@v5 + uses: actions/checkout@v6 with: fetch-depth: 0 ref: ${{ env.HEAD_REF }} diff --git a/.github/workflows/rollback-release.yml b/.github/workflows/rollback-release.yml index 8d8e872fa..8c9172021 100644 --- a/.github/workflows/rollback-release.yml +++ b/.github/workflows/rollback-release.yml @@ -52,7 +52,7 @@ jobs: steps: - name: Checkout repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 with: fetch-depth: 0 # Need full history for calculation of diffs @@ -137,7 +137,7 @@ jobs: - name: Generate token if: github.event_name == 'workflow_dispatch' - uses: actions/create-github-app-token@v2.1.4 + uses: actions/create-github-app-token@v2.2.0 id: app-token with: app-id: ${{ vars.AUTOMATION_APP_ID }} diff --git a/.github/workflows/test-codeql-bundle-all.yml b/.github/workflows/test-codeql-bundle-all.yml index 4b7fdca81..395288275 100644 --- a/.github/workflows/test-codeql-bundle-all.yml +++ b/.github/workflows/test-codeql-bundle-all.yml @@ -36,13 +36,17 @@ jobs: runs-on: ${{ matrix.os }} steps: - name: Check out repository - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Prepare test id: prepare-test uses: ./.github/actions/prepare-test with: version: ${{ matrix.version }} use-all-platform-bundle: true + - name: Install .NET + uses: actions/setup-dotnet@v5 + with: + dotnet-version: '9.x' - id: init uses: ./../action/init with: diff --git a/.github/workflows/update-bundle.yml b/.github/workflows/update-bundle.yml index 184c339ff..951b89066 100644 --- a/.github/workflows/update-bundle.yml +++ b/.github/workflows/update-bundle.yml @@ -33,7 +33,7 @@ jobs: GITHUB_CONTEXT: '${{ toJson(github) }}' run: echo "$GITHUB_CONTEXT" - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 - name: Update git config run: | diff --git a/.github/workflows/update-release-branch.yml b/.github/workflows/update-release-branch.yml index 830ed7c2a..9e196d2fe 100644 --- a/.github/workflows/update-release-branch.yml +++ b/.github/workflows/update-release-branch.yml @@ -38,7 +38,7 @@ jobs: contents: write # needed to push commits pull-requests: write # needed to create pull request steps: - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 with: fetch-depth: 0 # Need full history for calculation of diffs - uses: ./.github/actions/release-initialise @@ -93,14 +93,14 @@ jobs: pull-requests: write # needed to create pull request steps: - name: Generate token - uses: actions/create-github-app-token@v2.1.4 + uses: actions/create-github-app-token@v2.2.0 id: app-token with: app-id: ${{ vars.AUTOMATION_APP_ID }} private-key: ${{ secrets.AUTOMATION_PRIVATE_KEY }} - name: Checkout - uses: actions/checkout@v5 + uses: actions/checkout@v6 with: fetch-depth: 0 # Need full history for calculation of diffs token: ${{ steps.app-token.outputs.token }} diff --git a/.github/workflows/update-supported-enterprise-server-versions.yml b/.github/workflows/update-supported-enterprise-server-versions.yml index 421a63c69..4cead58f4 100644 --- a/.github/workflows/update-supported-enterprise-server-versions.yml +++ b/.github/workflows/update-supported-enterprise-server-versions.yml @@ -27,9 +27,9 @@ jobs: with: python-version: "3.13" - name: Checkout CodeQL Action - uses: actions/checkout@v5 + uses: actions/checkout@v6 - name: Checkout Enterprise Releases - uses: actions/checkout@v5 + uses: actions/checkout@v6 with: repository: github/enterprise-releases token: ${{ secrets.ENTERPRISE_RELEASE_TOKEN }} diff --git a/CHANGELOG.md b/CHANGELOG.md index b023f376b..1359cdfd9 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,7 +4,20 @@ See the [releases page](https://github.com/github/codeql-action/releases) for th ## [UNRELEASED] +No user facing changes. + +## 4.31.5 - 24 Nov 2025 + +- Update default CodeQL bundle version to 2.23.6. [#3321](https://github.com/github/codeql-action/pull/3321) + +## 4.31.4 - 18 Nov 2025 + +No user facing changes. + +## 4.31.3 - 13 Nov 2025 + - CodeQL Action v3 will be deprecated in December 2026. The Action now logs a warning for customers who are running v3 but could be running v4. For more information, see [Upcoming deprecation of CodeQL Action v3](https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/). +- Update default CodeQL bundle version to 2.23.5. [#3288](https://github.com/github/codeql-action/pull/3288) ## 4.31.2 - 30 Oct 2025 diff --git a/lib/analyze-action-post.js b/lib/analyze-action-post.js index 323549b8b..65c461c5a 100644 --- a/lib/analyze-action-post.js +++ b/lib/analyze-action-post.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug4("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay === "number" && delay > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve5(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug4; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning8(message, properties = {}) { + exports2.error = error3; + function warning9(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning8; + exports2.warning = warning9; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { @@ -27662,23 +27662,21 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -27686,10 +27684,10 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", @@ -27699,9 +27697,9 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27723,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -27911,8 +27910,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27921,8 @@ var require_light = __commonJS({ return (await Promise.all(promises3)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28034,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28071,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28089,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } } doExpire(clearGlobalState, run, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28116,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28395,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve5, returned, task; + var args, cb, error3, reject, resolve5, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve5, reject } = this._queue.shift()); @@ -28407,9 +28406,9 @@ var require_light = __commonJS({ return resolve5(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28542,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28876,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29179,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29212,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29237,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -33935,7 +33934,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core14 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob2 = __importStar4(require_glob()); var io6 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -34019,7 +34018,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core14.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -36085,8 +36084,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36320,9 +36319,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37653,14 +37652,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37670,7 +37669,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37678,8 +37677,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -37899,7 +37898,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39109,14 +39108,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39788,11 +39787,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39807,8 +39806,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40302,8 +40301,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40537,9 +40536,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41039,8 +41038,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41274,9 +41273,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42339,9 +42338,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42389,13 +42388,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42415,21 +42414,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42594,8 +42593,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43001,16 +43000,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43554,10 +43553,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45636,12 +45635,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -45906,16 +45905,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46040,7 +46039,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46206,7 +46205,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46447,9 +46446,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47157,7 +47156,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48220,26 +48219,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48280,12 +48279,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48293,13 +48292,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48308,7 +48307,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -62914,8 +62913,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -62949,10 +62948,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64536,8 +64535,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64552,9 +64551,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve5, reject) => { this.emitter.on("finish", resolve5); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65655,8 +65654,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65740,7 +65739,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -68892,7 +68891,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70259,9 +70258,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70375,12 +70374,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70414,13 +70413,13 @@ var require_requestUtils = __commonJS({ delay, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71236,8 +71235,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73083,8 +73082,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74655,12 +74654,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74680,12 +74679,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75149,8 +75148,8 @@ var require_test_transport = __commonJS({ } try { yield delay(this.responseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75161,8 +75160,8 @@ var require_test_transport = __commonJS({ stream.notifyMessage(msg); try { yield delay(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } } @@ -76225,8 +76224,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76322,8 +76321,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76354,18 +76353,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76633,8 +76632,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76835,22 +76834,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76905,15 +76904,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -76921,8 +76920,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76984,10 +76983,10 @@ var require_cache3 = __commonJS({ } core14.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else { @@ -77000,8 +76999,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77046,8 +77045,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core14.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core14.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError2(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core14.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77066,10 +77065,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77084,8 +77083,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79810,7 @@ var require_debug2 = __commonJS({ if (!debug4) { try { debug4 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug4 !== "function") { debug4 = function() { @@ -79844,8 +79843,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79918,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80087,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80289,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -83638,8 +83637,8 @@ var require_util11 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } function maskSecretUrls(body) { @@ -83732,8 +83731,8 @@ var require_artifact_twirp_client2 = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -83764,18 +83763,18 @@ var require_artifact_twirp_client2 = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -84046,11 +84045,11 @@ var require_blob_upload = __commonJS({ blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options), chunkTimer((0, config_1.getUploadChunkTimeout)()) ]); - } catch (error4) { - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + } catch (error3) { + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } - throw error4; + throw error3; } finally { abortController.abort(); } @@ -85071,9 +85070,9 @@ var require_async = __commonJS({ invokeCallback(callback, err && (err instanceof Error || err.message) ? err : new Error(err)); }); } - function invokeCallback(callback, error4, value) { + function invokeCallback(callback, error3, value) { try { - callback(error4, value); + callback(error3, value); } catch (err) { setImmediate$1((e) => { throw e; @@ -86379,10 +86378,10 @@ var require_async = __commonJS({ function reflect(fn) { var _fn = wrapAsync(fn); return initialParams(function reflectOn(args, reflectCallback) { - args.push((error4, ...cbArgs) => { + args.push((error3, ...cbArgs) => { let retVal = {}; - if (error4) { - retVal.error = error4; + if (error3) { + retVal.error = error3; } if (cbArgs.length > 0) { var value = cbArgs; @@ -86538,13 +86537,13 @@ var require_async = __commonJS({ var timer; function timeoutCallback() { var name = asyncFn.name || "anonymous"; - var error4 = new Error('Callback function "' + name + '" timed out.'); - error4.code = "ETIMEDOUT"; + var error3 = new Error('Callback function "' + name + '" timed out.'); + error3.code = "ETIMEDOUT"; if (info7) { - error4.info = info7; + error3.info = info7; } timedOut = true; - callback(error4); + callback(error3); } args.push((...cbArgs) => { if (!timedOut) { @@ -86587,7 +86586,7 @@ var require_async = __commonJS({ return callback[PROMISE_SYMBOL]; } function tryEach(tasks, callback) { - var error4 = null; + var error3 = null; var result; return eachSeries$1(tasks, (task, taskCb) => { wrapAsync(task)((err, ...args) => { @@ -86597,10 +86596,10 @@ var require_async = __commonJS({ } else { result = args; } - error4 = err; + error3 = err; taskCb(err ? null : {}); }); - }, () => callback(error4, result)); + }, () => callback(error3, result)); } var tryEach$1 = awaitify(tryEach); function unmemoize(fn) { @@ -93080,19 +93079,19 @@ var require_from = __commonJS({ next(); } }; - readable._destroy = function(error4, cb) { + readable._destroy = function(error3, cb) { PromisePrototypeThen( - close(error4), - () => process2.nextTick(cb, error4), + close(error3), + () => process2.nextTick(cb, error3), // nextTick is here in case cb throws - (e) => process2.nextTick(cb, e || error4) + (e) => process2.nextTick(cb, e || error3) ); }; - async function close(error4) { - const hadError = error4 !== void 0 && error4 !== null; + async function close(error3) { + const hadError = error3 !== void 0 && error3 !== null; const hasThrow = typeof iterator.throw === "function"; if (hadError && hasThrow) { - const { value, done } = await iterator.throw(error4); + const { value, done } = await iterator.throw(error3); await value; if (done) { return; @@ -93300,12 +93299,12 @@ var require_readable3 = __commonJS({ this.destroy(err); }; Readable.prototype[SymbolAsyncDispose] = function() { - let error4; + let error3; if (!this.destroyed) { - error4 = this.readableEnded ? null : new AbortError(); - this.destroy(error4); + error3 = this.readableEnded ? null : new AbortError(); + this.destroy(error3); } - return new Promise2((resolve5, reject) => eos(this, (err) => err && err !== error4 ? reject(err) : resolve5(null))); + return new Promise2((resolve5, reject) => eos(this, (err) => err && err !== error3 ? reject(err) : resolve5(null))); }; Readable.prototype.push = function(chunk, encoding) { return readableAddChunk(this, chunk, encoding, false); @@ -93858,14 +93857,14 @@ var require_readable3 = __commonJS({ } } stream.on("readable", next); - let error4; + let error3; const cleanup = eos( stream, { writable: false }, (err) => { - error4 = err ? aggregateTwoErrors(error4, err) : null; + error3 = err ? aggregateTwoErrors(error3, err) : null; callback(); callback = nop; } @@ -93875,19 +93874,19 @@ var require_readable3 = __commonJS({ const chunk = stream.destroyed ? null : stream.read(); if (chunk !== null) { yield chunk; - } else if (error4) { - throw error4; - } else if (error4 === null) { + } else if (error3) { + throw error3; + } else if (error3 === null) { return; } else { await new Promise2(next); } } } catch (err) { - error4 = aggregateTwoErrors(error4, err); - throw error4; + error3 = aggregateTwoErrors(error3, err); + throw error3; } finally { - if ((error4 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error4 === void 0 || stream._readableState.autoDestroy)) { + if ((error3 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error3 === void 0 || stream._readableState.autoDestroy)) { destroyImpl.destroyer(stream, null); } else { stream.off("readable", next); @@ -95380,11 +95379,11 @@ var require_pipeline3 = __commonJS({ yield* Readable.prototype[SymbolAsyncIterator].call(val2); } async function pumpToNode(iterable, writable, finish, { end }) { - let error4; + let error3; let onresolve = null; const resume = (err) => { if (err) { - error4 = err; + error3 = err; } if (onresolve) { const callback = onresolve; @@ -95393,12 +95392,12 @@ var require_pipeline3 = __commonJS({ } }; const wait = () => new Promise2((resolve5, reject) => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { onresolve = () => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve5(); } @@ -95428,7 +95427,7 @@ var require_pipeline3 = __commonJS({ } finish(); } catch (err) { - finish(error4 !== err ? aggregateTwoErrors(error4, err) : err); + finish(error3 !== err ? aggregateTwoErrors(error3, err) : err); } finally { cleanup(); writable.off("drain", resume); @@ -95482,7 +95481,7 @@ var require_pipeline3 = __commonJS({ if (outerSignal) { disposable = addAbortListener(outerSignal, abort); } - let error4; + let error3; let value; const destroys = []; let finishCount = 0; @@ -95491,23 +95490,23 @@ var require_pipeline3 = __commonJS({ } function finishImpl(err, final) { var _disposable; - if (err && (!error4 || error4.code === "ERR_STREAM_PREMATURE_CLOSE")) { - error4 = err; + if (err && (!error3 || error3.code === "ERR_STREAM_PREMATURE_CLOSE")) { + error3 = err; } - if (!error4 && !final) { + if (!error3 && !final) { return; } while (destroys.length) { - destroys.shift()(error4); + destroys.shift()(error3); } ; (_disposable = disposable) === null || _disposable === void 0 ? void 0 : _disposable[SymbolDispose](); ac.abort(); if (final) { - if (!error4) { + if (!error3) { lastStreamCleanup.forEach((fn) => fn()); } - process2.nextTick(callback, error4, value); + process2.nextTick(callback, error3, value); } } let ret; @@ -97357,52 +97356,128 @@ var require_isPlainObject = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/brace-expansion/index.js -var require_brace_expansion3 = __commonJS({ - "node_modules/archiver-utils/node_modules/brace-expansion/index.js"(exports2, module2) { - var balanced = require_balanced_match(); - module2.exports = expandTop; +// node_modules/@isaacs/balanced-match/dist/commonjs/index.js +var require_commonjs13 = __commonJS({ + "node_modules/@isaacs/balanced-match/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.range = exports2.balanced = void 0; + var balanced = (a, b, str2) => { + const ma = a instanceof RegExp ? maybeMatch(a, str2) : a; + const mb = b instanceof RegExp ? maybeMatch(b, str2) : b; + const r = ma !== null && mb != null && (0, exports2.range)(ma, mb, str2); + return r && { + start: r[0], + end: r[1], + pre: str2.slice(0, r[0]), + body: str2.slice(r[0] + ma.length, r[1]), + post: str2.slice(r[1] + mb.length) + }; + }; + exports2.balanced = balanced; + var maybeMatch = (reg, str2) => { + const m = str2.match(reg); + return m ? m[0] : null; + }; + var range = (a, b, str2) => { + let begs, beg, left, right = void 0, result; + let ai = str2.indexOf(a); + let bi = str2.indexOf(b, ai + 1); + let i = ai; + if (ai >= 0 && bi > 0) { + if (a === b) { + return [ai, bi]; + } + begs = []; + left = str2.length; + while (i >= 0 && !result) { + if (i === ai) { + begs.push(i); + ai = str2.indexOf(a, i + 1); + } else if (begs.length === 1) { + const r = begs.pop(); + if (r !== void 0) + result = [r, bi]; + } else { + beg = begs.pop(); + if (beg !== void 0 && beg < left) { + left = beg; + right = bi; + } + bi = str2.indexOf(b, i + 1); + } + i = ai < bi && ai >= 0 ? ai : bi; + } + if (begs.length && right !== void 0) { + result = [left, right]; + } + } + return result; + }; + exports2.range = range; + } +}); + +// node_modules/@isaacs/brace-expansion/dist/commonjs/index.js +var require_commonjs14 = __commonJS({ + "node_modules/@isaacs/brace-expansion/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.expand = expand; + var balanced_match_1 = require_commonjs13(); var escSlash = "\0SLASH" + Math.random() + "\0"; var escOpen = "\0OPEN" + Math.random() + "\0"; var escClose = "\0CLOSE" + Math.random() + "\0"; var escComma = "\0COMMA" + Math.random() + "\0"; var escPeriod = "\0PERIOD" + Math.random() + "\0"; + var escSlashPattern = new RegExp(escSlash, "g"); + var escOpenPattern = new RegExp(escOpen, "g"); + var escClosePattern = new RegExp(escClose, "g"); + var escCommaPattern = new RegExp(escComma, "g"); + var escPeriodPattern = new RegExp(escPeriod, "g"); + var slashPattern = /\\\\/g; + var openPattern = /\\{/g; + var closePattern = /\\}/g; + var commaPattern = /\\,/g; + var periodPattern = /\\./g; function numeric(str2) { - return parseInt(str2, 10) == str2 ? parseInt(str2, 10) : str2.charCodeAt(0); + return !isNaN(str2) ? parseInt(str2, 10) : str2.charCodeAt(0); } function escapeBraces(str2) { - return str2.split("\\\\").join(escSlash).split("\\{").join(escOpen).split("\\}").join(escClose).split("\\,").join(escComma).split("\\.").join(escPeriod); + return str2.replace(slashPattern, escSlash).replace(openPattern, escOpen).replace(closePattern, escClose).replace(commaPattern, escComma).replace(periodPattern, escPeriod); } function unescapeBraces(str2) { - return str2.split(escSlash).join("\\").split(escOpen).join("{").split(escClose).join("}").split(escComma).join(",").split(escPeriod).join("."); + return str2.replace(escSlashPattern, "\\").replace(escOpenPattern, "{").replace(escClosePattern, "}").replace(escCommaPattern, ",").replace(escPeriodPattern, "."); } function parseCommaParts(str2) { - if (!str2) + if (!str2) { return [""]; - var parts = []; - var m = balanced("{", "}", str2); - if (!m) + } + const parts = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) { return str2.split(","); - var pre = m.pre; - var body = m.body; - var post = m.post; - var p = pre.split(","); + } + const { pre, body, post } = m; + const p = pre.split(","); p[p.length - 1] += "{" + body + "}"; - var postParts = parseCommaParts(post); + const postParts = parseCommaParts(post); if (post.length) { + ; p[p.length - 1] += postParts.shift(); p.push.apply(p, postParts); } parts.push.apply(parts, p); return parts; } - function expandTop(str2) { - if (!str2) + function expand(str2) { + if (!str2) { return []; - if (str2.substr(0, 2) === "{}") { - str2 = "\\{\\}" + str2.substr(2); } - return expand(escapeBraces(str2), true).map(unescapeBraces); + if (str2.slice(0, 2) === "{}") { + str2 = "\\{\\}" + str2.slice(2); + } + return expand_(escapeBraces(str2), true).map(unescapeBraces); } function embrace(str2) { return "{" + str2 + "}"; @@ -97416,73 +97491,74 @@ var require_brace_expansion3 = __commonJS({ function gte5(i, y) { return i >= y; } - function expand(str2, isTop) { - var expansions = []; - var m = balanced("{", "}", str2); - if (!m) return [str2]; - var pre = m.pre; - var post = m.post.length ? expand(m.post, false) : [""]; + function expand_(str2, isTop) { + const expansions = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) + return [str2]; + const pre = m.pre; + const post = m.post.length ? expand_(m.post, false) : [""]; if (/\$$/.test(m.pre)) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + "{" + m.body + "}" + post[k]; + for (let k = 0; k < post.length; k++) { + const expansion = pre + "{" + m.body + "}" + post[k]; expansions.push(expansion); } } else { - var isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); - var isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); - var isSequence = isNumericSequence || isAlphaSequence; - var isOptions = m.body.indexOf(",") >= 0; + const isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); + const isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); + const isSequence = isNumericSequence || isAlphaSequence; + const isOptions = m.body.indexOf(",") >= 0; if (!isSequence && !isOptions) { if (m.post.match(/,(?!,).*\}/)) { str2 = m.pre + "{" + m.body + escClose + m.post; - return expand(str2); + return expand_(str2); } return [str2]; } - var n; + let n; if (isSequence) { n = m.body.split(/\.\./); } else { n = parseCommaParts(m.body); - if (n.length === 1) { - n = expand(n[0], false).map(embrace); + if (n.length === 1 && n[0] !== void 0) { + n = expand_(n[0], false).map(embrace); if (n.length === 1) { - return post.map(function(p) { - return m.pre + n[0] + p; - }); + return post.map((p) => m.pre + n[0] + p); } } } - var N; - if (isSequence) { - var x = numeric(n[0]); - var y = numeric(n[1]); - var width = Math.max(n[0].length, n[1].length); - var incr = n.length == 3 ? Math.abs(numeric(n[2])) : 1; - var test = lte; - var reverse = y < x; + let N; + if (isSequence && n[0] !== void 0 && n[1] !== void 0) { + const x = numeric(n[0]); + const y = numeric(n[1]); + const width = Math.max(n[0].length, n[1].length); + let incr = n.length === 3 && n[2] !== void 0 ? Math.abs(numeric(n[2])) : 1; + let test = lte; + const reverse = y < x; if (reverse) { incr *= -1; test = gte5; } - var pad = n.some(isPadded); + const pad = n.some(isPadded); N = []; - for (var i = x; test(i, y); i += incr) { - var c; + for (let i = x; test(i, y); i += incr) { + let c; if (isAlphaSequence) { c = String.fromCharCode(i); - if (c === "\\") + if (c === "\\") { c = ""; + } } else { c = String(i); if (pad) { - var need = width - c.length; + const need = width - c.length; if (need > 0) { - var z = new Array(need + 1).join("0"); - if (i < 0) + const z = new Array(need + 1).join("0"); + if (i < 0) { c = "-" + z + c.slice(1); - else + } else { c = z + c; + } } } } @@ -97490,15 +97566,16 @@ var require_brace_expansion3 = __commonJS({ } } else { N = []; - for (var j = 0; j < n.length; j++) { - N.push.apply(N, expand(n[j], false)); + for (let j = 0; j < n.length; j++) { + N.push.apply(N, expand_(n[j], false)); } } - for (var j = 0; j < N.length; j++) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + N[j] + post[k]; - if (!isTop || isSequence || expansion) + for (let j = 0; j < N.length; j++) { + for (let k = 0; k < post.length; k++) { + const expansion = pre + N[j] + post[k]; + if (!isTop || isSequence || expansion) { expansions.push(expansion); + } } } } @@ -97507,9 +97584,9 @@ var require_brace_expansion3 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js var require_assert_valid_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.assertValidPattern = void 0; @@ -97526,9 +97603,9 @@ var require_assert_valid_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js var require_brace_expressions = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.parseClass = void 0; @@ -97643,22 +97720,25 @@ var require_brace_expressions = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js var require_unescape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = void 0; - var unescape = (s, { windowsPathsNoEscape = false } = {}) => { - return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + var unescape = (s, { windowsPathsNoEscape = false, magicalBraces = true } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + } + return windowsPathsNoEscape ? s.replace(/\[([^\/\\{}])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\{}])\]/g, "$1$2").replace(/\\([^\/{}])/g, "$1"); }; exports2.unescape = unescape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js var require_ast = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.AST = void 0; @@ -98014,7 +98094,7 @@ var require_ast = __commonJS({ if (this.#root === this) this.#fillNegs(); if (!this.type) { - const noEmpty = this.isStart() && this.isEnd(); + const noEmpty = this.isStart() && this.isEnd() && !this.#parts.some((s) => typeof s !== "string"); const src = this.#parts.map((p) => { const [re, _2, hasMagic, uflag] = typeof p === "string" ? _AST.#parseGlob(p, this.#hasMagic, noEmpty) : p.toRegExpSource(allowDot); this.#hasMagic = this.#hasMagic || hasMagic; @@ -98124,10 +98204,7 @@ var require_ast = __commonJS({ } } if (c === "*") { - if (noEmpty && glob2 === "*") - re += starNoEmpty; - else - re += star; + re += noEmpty && glob2 === "*" ? starNoEmpty : star; hasMagic = true; continue; } @@ -98145,29 +98222,29 @@ var require_ast = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js var require_escape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.escape = void 0; - var escape = (s, { windowsPathsNoEscape = false } = {}) => { + var escape = (s, { windowsPathsNoEscape = false, magicalBraces = false } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/[?*()[\]{}]/g, "[$&]") : s.replace(/[?*()[\]\\{}]/g, "\\$&"); + } return windowsPathsNoEscape ? s.replace(/[?*()[\]]/g, "[$&]") : s.replace(/[?*()[\]\\]/g, "\\$&"); }; exports2.escape = escape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js -var require_commonjs13 = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js"(exports2) { +// node_modules/glob/node_modules/minimatch/dist/commonjs/index.js +var require_commonjs15 = __commonJS({ + "node_modules/glob/node_modules/minimatch/dist/commonjs/index.js"(exports2) { "use strict"; - var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { - return mod && mod.__esModule ? mod : { "default": mod }; - }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = exports2.escape = exports2.AST = exports2.Minimatch = exports2.match = exports2.makeRe = exports2.braceExpand = exports2.defaults = exports2.filter = exports2.GLOBSTAR = exports2.sep = exports2.minimatch = void 0; - var brace_expansion_1 = __importDefault4(require_brace_expansion3()); + var brace_expansion_1 = require_commonjs14(); var assert_valid_pattern_js_1 = require_assert_valid_pattern(); var ast_js_1 = require_ast(); var escape_js_1 = require_escape(); @@ -98290,7 +98367,7 @@ var require_commonjs13 = __commonJS({ if (options.nobrace || !/\{(?:(?!\{).)*\}/.test(pattern)) { return [pattern]; } - return (0, brace_expansion_1.default)(pattern); + return (0, brace_expansion_1.expand)(pattern); }; exports2.braceExpand = braceExpand; exports2.minimatch.braceExpand = exports2.braceExpand; @@ -98814,16 +98891,27 @@ var require_commonjs13 = __commonJS({ pp[i] = twoStar; } } else if (next === void 0) { - pp[i - 1] = prev + "(?:\\/|" + twoStar + ")?"; + pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + ")?"; } else if (next !== exports2.GLOBSTAR) { pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + "\\/)" + next; pp[i + 1] = exports2.GLOBSTAR; } }); - return pp.filter((p) => p !== exports2.GLOBSTAR).join("/"); + const filtered = pp.filter((p) => p !== exports2.GLOBSTAR); + if (this.partial && filtered.length >= 1) { + const prefixes = []; + for (let i = 1; i <= filtered.length; i++) { + prefixes.push(filtered.slice(0, i).join("/")); + } + return "(?:" + prefixes.join("|") + ")"; + } + return filtered.join("/"); }).join("|"); const [open, close] = set2.length > 1 ? ["(?:", ")"] : ["", ""]; re = "^" + open + re + close + "$"; + if (this.partial) { + re = "^(?:\\/|" + open + re.slice(1, -1) + close + ")$"; + } if (this.negate) re = "^(?!" + re + ").+$"; try { @@ -98910,9 +98998,9 @@ var require_commonjs13 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js -var require_commonjs14 = __commonJS({ - "node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js"(exports2) { +// node_modules/lru-cache/dist/commonjs/index.js +var require_commonjs16 = __commonJS({ + "node_modules/lru-cache/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.LRUCache = void 0; @@ -99001,6 +99089,7 @@ var require_commonjs14 = __commonJS({ #max; #maxSize; #dispose; + #onInsert; #disposeAfter; #fetchMethod; #memoMethod; @@ -99082,6 +99171,7 @@ var require_commonjs14 = __commonJS({ #hasDispose; #hasFetchMethod; #hasDisposeAfter; + #hasOnInsert; /** * Do not call this method unless you need to inspect the * inner workings of the cache. If anything returned by this @@ -99158,6 +99248,12 @@ var require_commonjs14 = __commonJS({ get dispose() { return this.#dispose; } + /** + * {@link LRUCache.OptionsBase.onInsert} (read-only) + */ + get onInsert() { + return this.#onInsert; + } /** * {@link LRUCache.OptionsBase.disposeAfter} (read-only) */ @@ -99165,7 +99261,7 @@ var require_commonjs14 = __commonJS({ return this.#disposeAfter; } constructor(options) { - const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; + const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, onInsert, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; if (max !== 0 && !isPosInt(max)) { throw new TypeError("max option must be a nonnegative integer"); } @@ -99207,6 +99303,9 @@ var require_commonjs14 = __commonJS({ if (typeof dispose === "function") { this.#dispose = dispose; } + if (typeof onInsert === "function") { + this.#onInsert = onInsert; + } if (typeof disposeAfter === "function") { this.#disposeAfter = disposeAfter; this.#disposed = []; @@ -99215,6 +99314,7 @@ var require_commonjs14 = __commonJS({ this.#disposed = void 0; } this.#hasDispose = !!this.#dispose; + this.#hasOnInsert = !!this.#onInsert; this.#hasDisposeAfter = !!this.#disposeAfter; this.noDisposeOnSet = !!noDisposeOnSet; this.noUpdateTTL = !!noUpdateTTL; @@ -99617,7 +99717,7 @@ var require_commonjs14 = __commonJS({ } /** * Return an array of [key, {@link LRUCache.Entry}] tuples which can be - * passed to {@link LRLUCache#load}. + * passed to {@link LRUCache#load}. * * The `start` fields are calculated relative to a portable `Date.now()` * timestamp, even if `performance.now()` is available. @@ -99728,6 +99828,9 @@ var require_commonjs14 = __commonJS({ if (status) status.set = "add"; noUpdateTTL = false; + if (this.#hasOnInsert) { + this.#onInsert?.(v, k, "add"); + } } else { this.#moveToTail(index); const oldVal = this.#valList[index]; @@ -99763,6 +99866,9 @@ var require_commonjs14 = __commonJS({ } else if (status) { status.set = "update"; } + if (this.#hasOnInsert) { + this.onInsert?.(v, k, v === oldVal ? "update" : "replace"); + } } if (ttl !== 0 && !this.#ttls) { this.#initializeTTLTracking(); @@ -100288,7 +100394,7 @@ var require_commonjs14 = __commonJS({ }); // node_modules/minipass/dist/commonjs/index.js -var require_commonjs15 = __commonJS({ +var require_commonjs17 = __commonJS({ "node_modules/minipass/dist/commonjs/index.js"(exports2) { "use strict"; var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { @@ -101179,9 +101285,9 @@ var require_commonjs15 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js -var require_commonjs16 = __commonJS({ - "node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js"(exports2) { +// node_modules/path-scurry/dist/commonjs/index.js +var require_commonjs18 = __commonJS({ + "node_modules/path-scurry/dist/commonjs/index.js"(exports2) { "use strict"; var __createBinding4 = exports2 && exports2.__createBinding || (Object.create ? (function(o, m, k, k2) { if (k2 === void 0) k2 = k; @@ -101212,14 +101318,14 @@ var require_commonjs16 = __commonJS({ }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.PathScurry = exports2.Path = exports2.PathScurryDarwin = exports2.PathScurryPosix = exports2.PathScurryWin32 = exports2.PathScurryBase = exports2.PathPosix = exports2.PathWin32 = exports2.PathBase = exports2.ChildrenCache = exports2.ResolveCache = void 0; - var lru_cache_1 = require_commonjs14(); + var lru_cache_1 = require_commonjs16(); var node_path_1 = require("node:path"); var node_url_1 = require("node:url"); var fs_1 = require("fs"); var actualFS = __importStar4(require("node:fs")); var realpathSync = fs_1.realpathSync.native; var promises_1 = require("node:fs/promises"); - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var defaultFS = { lstatSync: fs_1.lstatSync, readdir: fs_1.readdir, @@ -101434,6 +101540,8 @@ var require_commonjs16 = __commonJS({ /** * Deprecated alias for Dirent['parentPath'] Somewhat counterintuitively, * this property refers to the *parent* path, not the path object itself. + * + * @deprecated */ get path() { return this.parentPath; @@ -102953,13 +103061,13 @@ var require_commonjs16 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js +// node_modules/glob/dist/commonjs/pattern.js var require_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js"(exports2) { + "node_modules/glob/dist/commonjs/pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Pattern = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var isPatternList = (pl) => pl.length >= 1; var isGlobList = (gl) => gl.length >= 1; var Pattern = class _Pattern { @@ -103127,13 +103235,13 @@ var require_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js +// node_modules/glob/dist/commonjs/ignore.js var require_ignore = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js"(exports2) { + "node_modules/glob/dist/commonjs/ignore.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Ignore = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var pattern_js_1 = require_pattern(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; var Ignore = class { @@ -103224,13 +103332,13 @@ var require_ignore = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js +// node_modules/glob/dist/commonjs/processor.js var require_processor = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js"(exports2) { + "node_modules/glob/dist/commonjs/processor.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Processor = exports2.SubWalks = exports2.MatchRecord = exports2.HasWalkedCache = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var HasWalkedCache = class _HasWalkedCache { store; constructor(store = /* @__PURE__ */ new Map()) { @@ -103457,13 +103565,13 @@ var require_processor = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js +// node_modules/glob/dist/commonjs/walker.js var require_walker = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js"(exports2) { + "node_modules/glob/dist/commonjs/walker.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.GlobStream = exports2.GlobWalker = exports2.GlobUtil = void 0; - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var ignore_js_1 = require_ignore(); var processor_js_1 = require_processor(); var makeIgnore = (ignore, opts) => typeof ignore === "string" ? new ignore_js_1.Ignore([ignore], opts) : Array.isArray(ignore) ? new ignore_js_1.Ignore(ignore, opts) : ignore; @@ -103797,15 +103905,15 @@ var require_walker = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js +// node_modules/glob/dist/commonjs/glob.js var require_glob2 = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js"(exports2) { + "node_modules/glob/dist/commonjs/glob.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Glob = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var node_url_1 = require("node:url"); - var path_scurry_1 = require_commonjs16(); + var path_scurry_1 = require_commonjs18(); var pattern_js_1 = require_pattern(); var walker_js_1 = require_walker(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; @@ -104010,13 +104118,13 @@ var require_glob2 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js +// node_modules/glob/dist/commonjs/has-magic.js var require_has_magic = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js"(exports2) { + "node_modules/glob/dist/commonjs/has-magic.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.hasMagic = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var hasMagic = (pattern, options = {}) => { if (!Array.isArray(pattern)) { pattern = [pattern]; @@ -104031,9 +104139,9 @@ var require_has_magic = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js -var require_commonjs17 = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js"(exports2) { +// node_modules/glob/dist/commonjs/index.js +var require_commonjs19 = __commonJS({ + "node_modules/glob/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.glob = exports2.sync = exports2.iterate = exports2.iterateSync = exports2.stream = exports2.streamSync = exports2.Ignore = exports2.hasMagic = exports2.Glob = exports2.unescape = exports2.escape = void 0; @@ -104042,10 +104150,10 @@ var require_commonjs17 = __commonJS({ exports2.globSync = globSync; exports2.globIterateSync = globIterateSync; exports2.globIterate = globIterate; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var glob_js_1 = require_glob2(); var has_magic_js_1 = require_has_magic(); - var minimatch_2 = require_commonjs13(); + var minimatch_2 = require_commonjs15(); Object.defineProperty(exports2, "escape", { enumerable: true, get: function() { return minimatch_2.escape; } }); @@ -104122,7 +104230,7 @@ var require_file3 = __commonJS({ var difference = require_difference(); var union = require_union(); var isPlainObject = require_isPlainObject(); - var glob2 = require_commonjs17(); + var glob2 = require_commonjs19(); var file = module2.exports = {}; var pathSeparatorRe = /[\/\\]/g; var processPatterns = function(patterns, fn) { @@ -105849,18 +105957,18 @@ var require_zip_archive_output_stream = __commonJS({ ZipArchiveOutputStream.prototype._smartStream = function(ae, callback) { var deflate = ae.getMethod() === constants.METHOD_DEFLATED; var process2 = deflate ? new DeflateCRC32Stream(this.options.zlib) : new CRC32Stream(); - var error4 = null; + var error3 = null; function handleStuff() { var digest = process2.digest().readUInt32BE(0); ae.setCrc(digest); ae.setSize(process2.size()); ae.setCompressedSize(process2.size(true)); this._afterAppend(ae); - callback(error4, ae); + callback(error3, ae); } process2.once("end", handleStuff.bind(this)); process2.once("error", function(err) { - error4 = err; + error3 = err; }); process2.pipe(this, { end: false }); return process2; @@ -107226,11 +107334,11 @@ var require_streamx = __commonJS({ } [asyncIterator]() { const stream = this; - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("readable", onreadable); this.on("close", onclose); @@ -107262,7 +107370,7 @@ var require_streamx = __commonJS({ } function ondata(data) { if (promiseReject === null) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else if (data === null && (stream._duplexState & READ_DONE) === 0) promiseReject(STREAM_DESTROYED); else promiseResolve({ value: data, done: data === null }); promiseReject = promiseResolve = null; @@ -107436,7 +107544,7 @@ var require_streamx = __commonJS({ if (all.length < 2) throw new Error("Pipeline requires at least 2 streams"); let src = all[0]; let dest = null; - let error4 = null; + let error3 = null; for (let i = 1; i < all.length; i++) { dest = all[i]; if (isStreamx(src)) { @@ -107451,14 +107559,14 @@ var require_streamx = __commonJS({ let fin = false; const autoDestroy = isStreamx(dest) || !!(dest._writableState && dest._writableState.autoDestroy); dest.on("error", (err) => { - if (error4 === null) error4 = err; + if (error3 === null) error3 = err; }); dest.on("finish", () => { fin = true; - if (!autoDestroy) done(error4); + if (!autoDestroy) done(error3); }); if (autoDestroy) { - dest.on("close", () => done(error4 || (fin ? null : PREMATURE_CLOSE))); + dest.on("close", () => done(error3 || (fin ? null : PREMATURE_CLOSE))); } } return dest; @@ -107471,8 +107579,8 @@ var require_streamx = __commonJS({ } } function onerror(err) { - if (!err || error4) return; - error4 = err; + if (!err || error3) return; + error3 = err; for (const s of all) { s.destroy(err); } @@ -108038,7 +108146,7 @@ var require_extract = __commonJS({ cb(null); } [Symbol.asyncIterator]() { - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; let entryStream = null; @@ -108046,7 +108154,7 @@ var require_extract = __commonJS({ const extract2 = this; this.on("entry", onentry); this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("close", onclose); return { @@ -108070,8 +108178,8 @@ var require_extract = __commonJS({ cb(err); } function onnext(resolve5, reject) { - if (error4) { - return reject(error4); + if (error3) { + return reject(error3); } if (entryStream) { resolve5({ value: entryStream, done: false }); @@ -108097,9 +108205,9 @@ var require_extract = __commonJS({ } } function onclose() { - consumeCallback(error4); + consumeCallback(error3); if (!promiseResolve) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else promiseResolve({ value: void 0, done: true }); promiseResolve = promiseReject = null; } @@ -108995,18 +109103,18 @@ var require_zip2 = __commonJS({ return zipUploadStream; }); } - var zipErrorCallback = (error4) => { + var zipErrorCallback = (error3) => { core14.error("An error has occurred while creating the zip file for upload"); - core14.info(error4); + core14.info(error3); throw new Error("An error has occurred during zip creation for the artifact"); }; - var zipWarningCallback = (error4) => { - if (error4.code === "ENOENT") { + var zipWarningCallback = (error3) => { + if (error3.code === "ENOENT") { core14.warning("ENOENT warning during artifact zip creation. No such file or directory"); - core14.info(error4); + core14.info(error3); } else { - core14.warning(`A non-blocking warning has occurred during artifact zip creation: ${error4.code}`); - core14.info(error4); + core14.warning(`A non-blocking warning has occurred during artifact zip creation: ${error3.code}`); + core14.info(error3); } }; var zipFinishCallback = () => { @@ -110813,8 +110921,8 @@ var require_parser_stream = __commonJS({ this.unzipStream.on("entry", function(entry) { self2.push(entry); }); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(ParserStream, Transform); @@ -110954,8 +111062,8 @@ var require_extract2 = __commonJS({ this.createdDirectories = {}; var self2 = this; this.unzipStream.on("entry", this._processEntry.bind(this)); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(Extract, Transform); @@ -110989,8 +111097,8 @@ var require_extract2 = __commonJS({ self2.unfinishedEntries--; self2._notifyAwaiter(); }); - pipedStream.on("error", function(error4) { - self2.emit("error", error4); + pipedStream.on("error", function(error3) { + self2.emit("error", error3); }); entry.pipe(pipedStream); }; @@ -111125,11 +111233,11 @@ var require_download_artifact = __commonJS({ try { yield promises_1.default.access(path6); return true; - } catch (error4) { - if (error4.code === "ENOENT") { + } catch (error3) { + if (error3.code === "ENOENT") { return false; } else { - throw error4; + throw error3; } } }); @@ -111140,9 +111248,9 @@ var require_download_artifact = __commonJS({ while (retryCount < 5) { try { return yield streamExtractExternal(url, directory); - } catch (error4) { + } catch (error3) { retryCount++; - core14.debug(`Failed to download artifact after ${retryCount} retries due to ${error4.message}. Retrying in 5 seconds...`); + core14.debug(`Failed to download artifact after ${retryCount} retries due to ${error3.message}. Retrying in 5 seconds...`); yield new Promise((resolve5) => setTimeout(resolve5, 5e3)); } } @@ -111171,10 +111279,10 @@ var require_download_artifact = __commonJS({ const extractStream = passThrough; extractStream.on("data", () => { timer.refresh(); - }).on("error", (error4) => { - core14.debug(`response.message: Artifact download failed: ${error4.message}`); + }).on("error", (error3) => { + core14.debug(`response.message: Artifact download failed: ${error3.message}`); clearTimeout(timer); - reject(error4); + reject(error3); }).pipe(unzip_stream_1.default.Extract({ path: directory })).on("close", () => { clearTimeout(timer); if (hashStream) { @@ -111183,8 +111291,8 @@ var require_download_artifact = __commonJS({ core14.info(`SHA256 digest of downloaded artifact is ${sha256Digest}`); } resolve5({ sha256Digest: `sha256:${sha256Digest}` }); - }).on("error", (error4) => { - reject(error4); + }).on("error", (error3) => { + reject(error3); }); }); }); @@ -111223,8 +111331,8 @@ var require_download_artifact = __commonJS({ core14.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -111266,8 +111374,8 @@ Are you trying to download from a different run? Try specifying a github-token w core14.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -111365,9 +111473,9 @@ var require_dist_node16 = __commonJS({ return request(options).then((response) => { octokit.log.info(`${requestOptions.method} ${path6} - ${response.status} in ${Date.now() - start}ms`); return response; - }).catch((error4) => { - octokit.log.info(`${requestOptions.method} ${path6} - ${error4.status} in ${Date.now() - start}ms`); - throw error4; + }).catch((error3) => { + octokit.log.info(`${requestOptions.method} ${path6} - ${error3.status} in ${Date.now() - start}ms`); + throw error3; }); }); } @@ -111385,22 +111493,22 @@ var require_dist_node17 = __commonJS({ return ex && typeof ex === "object" && "default" in ex ? ex["default"] : ex; } var Bottleneck = _interopDefault(require_light()); - async function errorRequest(octokit, state, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(octokit, state, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } async function wrapRequest(state, request, options) { const limiter = new Bottleneck(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -111422,12 +111530,12 @@ var require_dist_node17 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -111918,13 +112026,13 @@ var require_client2 = __commonJS({ throw new errors_1.GHESNotSupportedError(); } return (0, upload_artifact_1.uploadArtifact)(name, files, rootDirectory, options); - } catch (error4) { - (0, core_1.warning)(`Artifact upload failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Artifact upload failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions is operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -111939,13 +112047,13 @@ If the error persists, please check whether Actions is operating normally at [ht return (0, download_artifact_1.downloadArtifactPublic)(artifactId, repositoryOwner, repositoryName, token, downloadOptions); } return (0, download_artifact_1.downloadArtifactInternal)(artifactId, options); - } catch (error4) { - (0, core_1.warning)(`Download Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Download Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -111960,13 +112068,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, list_artifacts_1.listArtifactsPublic)(workflowRunId, repositoryOwner, repositoryName, token, options === null || options === void 0 ? void 0 : options.latest); } return (0, list_artifacts_1.listArtifactsInternal)(options === null || options === void 0 ? void 0 : options.latest); - } catch (error4) { - (0, core_1.warning)(`Listing Artifacts failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Listing Artifacts failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -111981,13 +112089,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, get_artifact_1.getArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, get_artifact_1.getArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Get Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Get Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -112002,13 +112110,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, delete_artifact_1.deleteArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, delete_artifact_1.deleteArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Delete Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Delete Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -112508,14 +112616,14 @@ var require_tmp = __commonJS({ options.template = _getRelativePathSync("template", options.template, tmpDir); return options; } - function _isEBADF(error4) { - return _isExpectedError(error4, -EBADF, "EBADF"); + function _isEBADF(error3) { + return _isExpectedError(error3, -EBADF, "EBADF"); } - function _isENOENT(error4) { - return _isExpectedError(error4, -ENOENT, "ENOENT"); + function _isENOENT(error3) { + return _isExpectedError(error3, -ENOENT, "ENOENT"); } - function _isExpectedError(error4, errno, code) { - return IS_WIN32 ? error4.code === code : error4.code === code && error4.errno === errno; + function _isExpectedError(error3, errno, code) { + return IS_WIN32 ? error3.code === code : error3.code === code && error3.errno === errno; } function setGracefulCleanup() { _gracefulCleanup = true; @@ -114227,9 +114335,9 @@ var require_upload_gzip = __commonJS({ const size = (yield stat(tempFilePath)).size; resolve5(size); })); - outputStream.on("error", (error4) => { - console.log(error4); - reject(error4); + outputStream.on("error", (error3) => { + console.log(error3); + reject(error3); }); }); }); @@ -114354,9 +114462,9 @@ var require_requestUtils2 = __commonJS({ } isRetryable = (0, utils_1.isRetryableStatusCode)(statusCode); errorMessage = `Artifact service responded with ${statusCode}`; - } catch (error4) { + } catch (error3) { isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { core14.info(`${name} - Error is not retryable`); @@ -114722,9 +114830,9 @@ var require_upload_http_client = __commonJS({ let response; try { response = yield uploadChunkRequest(); - } catch (error4) { + } catch (error3) { core14.info(`An error has been caught http-client index ${httpClientIndex}, retrying the upload`); - console.log(error4); + console.log(error3); if (incrementAndCheckRetryLimit()) { return false; } @@ -114913,8 +115021,8 @@ var require_download_http_client = __commonJS({ } this.statusReporter.incrementProcessedCount(); } - }))).catch((error4) => { - throw new Error(`Unable to download the artifact: ${error4}`); + }))).catch((error3) => { + throw new Error(`Unable to download the artifact: ${error3}`); }).finally(() => { this.statusReporter.stop(); this.downloadHttpManager.disposeAndReplaceAllClients(); @@ -114979,9 +115087,9 @@ var require_download_http_client = __commonJS({ let response; try { response = yield makeDownloadRequest(); - } catch (error4) { + } catch (error3) { core14.info("An error occurred while attempting to download a file"); - console.log(error4); + console.log(error3); yield backOff(); continue; } @@ -114995,7 +115103,7 @@ var require_download_http_client = __commonJS({ } else { forceRetry = true; } - } catch (error4) { + } catch (error3) { forceRetry = true; } } @@ -115021,31 +115129,31 @@ var require_download_http_client = __commonJS({ yield new Promise((resolve5, reject) => { if (isGzip) { const gunzip = zlib.createGunzip(); - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core14.info(`An error occurred while attempting to read the response stream`); gunzip.close(); destinationStream.close(); - reject(error4); - }).pipe(gunzip).on("error", (error4) => { + reject(error3); + }).pipe(gunzip).on("error", (error3) => { core14.info(`An error occurred while attempting to decompress the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve5(); - }).on("error", (error4) => { + }).on("error", (error3) => { core14.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } else { - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core14.info(`An error occurred while attempting to read the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve5(); - }).on("error", (error4) => { + }).on("error", (error3) => { core14.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } }); @@ -116446,7 +116554,6 @@ var io2 = __toESM(require_io2()); // src/util.ts var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js @@ -116464,14 +116571,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs7.lstat(itemPath, { bigint: true }) : await fs7.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs7.lstat(itemPath, { bigint: true }) : await fs7.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs7.readdir(itemPath) : await fs7.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs7.readdir(itemPath) : await fs7.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -116482,13 +116589,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -117368,6 +117475,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -117487,7 +117606,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -117527,16 +117646,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -119106,9 +119216,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -119201,11 +119311,11 @@ function getCachedCodeQlVersion() { async function codeQlVersionAtLeast(codeql, requiredVersion) { return semver.gte((await codeql.getVersion()).version, requiredVersion); } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } function cloneObject(obj) { return JSON.parse(JSON.stringify(obj)); @@ -119406,19 +119516,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -119434,9 +119544,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -119671,13 +119781,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -119835,7 +119945,7 @@ function withGroup(groupName, f) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { @@ -119966,6 +120076,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -120077,21 +120192,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, @@ -120104,6 +120219,8 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, @@ -120899,16 +121016,16 @@ async function runWrapper() { if (fs6.existsSync(tempDependencyDir)) { try { fs6.rmSync(tempDependencyDir, { recursive: true }); - } catch (error4) { + } catch (error3) { logger.info( - `Failed to remove temporary dependencies directory: ${getErrorMessage(error4)}` + `Failed to remove temporary dependencies directory: ${getErrorMessage(error3)}` ); } } } - } catch (error4) { + } catch (error3) { core13.setFailed( - `analyze post-action step failed: ${getErrorMessage(error4)}` + `analyze post-action step failed: ${getErrorMessage(error3)}` ); } } @@ -120997,5 +121114,5 @@ tmp/lib/tmp.js: *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/analyze-action.js b/lib/analyze-action.js index e5e897ca3..0c9e35dcb 100644 --- a/lib/analyze-action.js +++ b/lib/analyze-action.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug5("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve8(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug5; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning9(message, properties = {}) { + exports2.error = error3; + function warning10(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning9; + exports2.warning = warning10; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url2 = ""; return { value: { @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { @@ -27662,23 +27662,21 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -27686,10 +27684,10 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", @@ -27699,9 +27697,9 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27723,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -27911,8 +27910,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27921,8 @@ var require_light = __commonJS({ return (await Promise.all(promises4)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28034,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28071,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run2, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28089,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } } doExpire(clearGlobalState, run2, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run2, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run2, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28116,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28395,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve8, returned, task; + var args, cb, error3, reject, resolve8, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve8, reject } = this._queue.shift()); @@ -28407,9 +28406,9 @@ var require_light = __commonJS({ return resolve8(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28542,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28876,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29179,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29212,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29237,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -33935,7 +33934,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core15 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob2 = __importStar4(require_glob()); var io7 = __importStar4(require_io3()); var crypto2 = __importStar4(require("crypto")); @@ -34019,7 +34018,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core15.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -36085,8 +36084,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36320,9 +36319,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37653,14 +37652,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37670,7 +37669,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37678,8 +37677,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -37899,7 +37898,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39109,14 +39108,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39788,11 +39787,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39807,8 +39806,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40302,8 +40301,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40537,9 +40536,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41039,8 +41038,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41274,9 +41273,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42339,9 +42338,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42389,13 +42388,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42415,21 +42414,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42594,8 +42593,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43001,16 +43000,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43554,10 +43553,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45636,12 +45635,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -45906,16 +45905,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46040,7 +46039,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46206,7 +46205,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46447,9 +46446,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47157,7 +47156,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48220,26 +48219,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48280,12 +48279,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48293,13 +48292,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48308,7 +48307,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -62914,8 +62913,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -62949,10 +62948,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64536,8 +64535,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64552,9 +64551,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve8, reject) => { this.emitter.on("finish", resolve8); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65655,8 +65654,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65740,7 +65739,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -68892,7 +68891,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70259,9 +70258,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core15.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core15.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70375,12 +70374,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70414,13 +70413,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71236,8 +71235,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73083,8 +73082,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74655,12 +74654,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74680,12 +74679,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75149,8 +75148,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75161,8 +75160,8 @@ var require_test_transport = __commonJS({ stream2.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } } @@ -76225,8 +76224,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url2} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url2} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76322,8 +76321,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url2, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76354,18 +76353,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76633,8 +76632,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76835,22 +76834,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core15.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core15.error(`Failed to restore: ${error4.message}`); + core15.error(`Failed to restore: ${error3.message}`); } else { - core15.warning(`Failed to restore: ${error4.message}`); + core15.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core15.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core15.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76905,15 +76904,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core15.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core15.error(`Failed to restore: ${error4.message}`); + core15.error(`Failed to restore: ${error3.message}`); } else { - core15.warning(`Failed to restore: ${error4.message}`); + core15.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -76921,8 +76920,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core15.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core15.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76984,10 +76983,10 @@ var require_cache3 = __commonJS({ } core15.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core15.info(`Failed to save: ${typedError.message}`); } else { @@ -77000,8 +76999,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core15.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core15.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77046,8 +77045,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core15.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core15.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError2(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core15.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77066,10 +77065,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core15.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77084,8 +77083,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core15.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core15.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79810,7 @@ var require_debug2 = __commonJS({ if (!debug5) { try { debug5 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug5 !== "function") { debug5 = function() { @@ -79844,8 +79843,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79918,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80087,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80289,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -84337,7 +84336,6 @@ var fsPromises = __toESM(require("fs/promises")); var os = __toESM(require("os")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js @@ -84355,14 +84353,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs17.lstat(itemPath, { bigint: true }) : await fs17.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs17.lstat(itemPath, { bigint: true }) : await fs17.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs17.readdir(itemPath) : await fs17.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs17.readdir(itemPath) : await fs17.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -84373,13 +84371,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -85259,6 +85257,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -85378,7 +85388,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -85418,16 +85428,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -86993,9 +86994,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -87379,11 +87380,11 @@ function parseMatrixInput(matrixInput) { } return JSON.parse(matrixInput); } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } async function checkDiskUsage(logger) { try { @@ -87406,9 +87407,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -87420,7 +87421,7 @@ function checkActionVersion(version, githubVersion) { semver.coerce(githubVersion.version) ?? "0.0.0", ">=3.20" )) { - core3.error( + core3.warning( "CodeQL Action v3 will be deprecated in December 2026. Please update all occurrences of the CodeQL Action in your workflow files to v4. For more information, see https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/" ); core3.exportVariable("CODEQL_ACTION_DID_LOG_VERSION_DEPRECATION" /* LOG_VERSION_DEPRECATION */, "true"); @@ -87958,19 +87959,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -87986,9 +87987,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -88210,8 +88211,8 @@ var path4 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); @@ -88240,13 +88241,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -88449,7 +88450,7 @@ function formatDuration(durationMs) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { @@ -88521,7 +88522,7 @@ function checkOverlayBaseDatabase(config, logger, warningPrefix) { } return true; } -async function uploadOverlayBaseDatabaseToCache(codeql, config, logger) { +async function cleanupAndUploadOverlayBaseDatabaseToCache(codeql, config, logger) { const overlayDatabaseMode = config.overlayDatabaseMode; if (overlayDatabaseMode !== "overlay-base" /* OverlayBase */) { logger.debug( @@ -88550,7 +88551,7 @@ async function uploadOverlayBaseDatabaseToCache(codeql, config, logger) { return false; } await withGroupAsync("Cleaning up databases", async () => { - await codeql.databaseCleanupCluster(config, "overlay"); + await codeql.databaseCleanupCluster(config, "overlay" /* Overlay */); }); const dbLocation = config.dbLocation; const databaseSizeBytes = await tryGetFolderBytes(dbLocation, logger); @@ -88589,9 +88590,9 @@ async function uploadOverlayBaseDatabaseToCache(codeql, config, logger) { logger.warning("Timed out while uploading overlay-base database"); return false; } - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to upload overlay-base database to cache: ${error4 instanceof Error ? error4.message : String(error4)}` + `Failed to upload overlay-base database to cache: ${error3 instanceof Error ? error3.message : String(error3)}` ); return false; } @@ -88696,6 +88697,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -88807,21 +88813,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, @@ -89178,17 +89184,17 @@ async function getFileDiffsWithBasehead(branches, logger) { ${JSON.stringify(response, null, 2)}` ); return response.data.files; - } catch (error4) { - if (error4.status) { - logger.warning(`Error retrieving diff ${basehead}: ${error4.message}`); + } catch (error3) { + if (error3.status) { + logger.warning(`Error retrieving diff ${basehead}: ${error3.message}`); logger.debug( `Error running compareCommitsWithBasehead(${basehead}): -Request: ${JSON.stringify(error4.request, null, 2)} -Error Response: ${JSON.stringify(error4.response, null, 2)}` +Request: ${JSON.stringify(error3.request, null, 2)} +Error Response: ${JSON.stringify(error3.response, null, 2)}` ); return void 0; } else { - throw error4; + throw error3; } } } @@ -89372,6 +89378,8 @@ async function cachePrefix(codeql, language) { } // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, @@ -91173,6 +91181,11 @@ async function uploadDependencyCaches(codeql, features, config, logger) { status.push({ language, result: "no-hash" /* NoHash */ }); continue; } + const key = await cacheKey2(codeql, features, language, patterns); + if (config.dependencyCachingRestoredKeys.includes(key)) { + status.push({ language, result: "duplicate" /* Duplicate */ }); + continue; + } const size = await getTotalCacheSize( await cacheConfig.getDependencyPaths(codeql, features), logger, @@ -91185,7 +91198,6 @@ async function uploadDependencyCaches(codeql, features, config, logger) { ); continue; } - const key = await cacheKey2(codeql, features, language, patterns); logger.info( `Uploading cache of size ${size} for ${language} with key ${key}...` ); @@ -91202,15 +91214,15 @@ async function uploadDependencyCaches(codeql, features, config, logger) { upload_size_bytes: Math.round(size), upload_duration_ms }); - } catch (error4) { - if (error4 instanceof actionsCache3.ReserveCacheError) { + } catch (error3) { + if (error3 instanceof actionsCache3.ReserveCacheError) { logger.info( `Not uploading cache for ${language}, because ${key} is already in use.` ); - logger.debug(error4.message); + logger.debug(error3.message); status.push({ language, result: "duplicate" /* Duplicate */ }); } else { - throw error4; + throw error3; } } } @@ -91309,11 +91321,11 @@ function writeDiagnostic(config, language, diagnostic) { // src/analyze.ts var CodeQLAnalysisError = class extends Error { - constructor(queriesStatusReport, message, error4) { + constructor(queriesStatusReport, message, error3) { super(message); this.queriesStatusReport = queriesStatusReport; this.message = message; - this.error = error4; + this.error = error3; this.name = "CodeQLAnalysisError"; } }; @@ -91635,9 +91647,9 @@ async function runQueries(sarifFolder, memoryFlag, threadsFlag, diffRangePackDir async function runFinalize(features, outputDir, threadsFlag, memoryFlag, codeql, config, logger) { try { await fs12.promises.rm(outputDir, { force: true, recursive: true }); - } catch (error4) { - if (error4?.code !== "ENOENT") { - throw error4; + } catch (error3) { + if (error3?.code !== "ENOENT") { + throw error3; } } await fs12.promises.mkdir(outputDir, { recursive: true }); @@ -91685,7 +91697,7 @@ async function warnIfGoInstalledAfterInit(config, logger) { // src/database-upload.ts var fs13 = __toESM(require("fs")); -async function uploadDatabases(repositoryNwo, codeql, config, apiDetails, logger) { +async function cleanupAndUploadDatabases(repositoryNwo, codeql, config, apiDetails, features, logger) { if (getRequiredInput("upload-database") !== "true") { logger.debug("Database upload disabled in workflow. Skipping upload."); return; @@ -91708,8 +91720,9 @@ async function uploadDatabases(repositoryNwo, codeql, config, apiDetails, logger logger.debug("Not analyzing default branch. Skipping upload."); return; } + const cleanupLevel = config.overlayDatabaseMode === "overlay-base" /* OverlayBase */ && await features.getValue("upload_overlay_db_to_api" /* UploadOverlayDbToApi */) ? "overlay" /* Overlay */ : "clear" /* Clear */; await withGroupAsync("Cleaning up databases", async () => { - await codeql.databaseCleanupCluster(config, "clear"); + await codeql.databaseCleanupCluster(config, cleanupLevel); }); const client = getApiClient(); const uploadsUrl = new URL(parseGitHubUrl(apiDetails.url)); @@ -91763,9 +91776,9 @@ function isFirstPartyAnalysis(actionName) { } return process.env["CODEQL_ACTION_INIT_HAS_RUN" /* INIT_ACTION_HAS_RUN */] === "true"; } -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -93432,15 +93445,15 @@ function validateSarifFileSchema(sarif, sarifFilePath, logger) { const warnings = (result.errors ?? []).filter( (err) => err.name === "format" && typeof err.argument === "string" && warningAttributes.includes(err.argument) ); - for (const warning9 of warnings) { + for (const warning10 of warnings) { logger.info( - `Warning: '${warning9.instance}' is not a valid URI in '${warning9.property}'.` + `Warning: '${warning10.instance}' is not a valid URI in '${warning10.property}'.` ); } if (errors.length > 0) { - for (const error4 of errors) { - logger.startGroup(`Error details: ${error4.stack}`); - logger.info(JSON.stringify(error4, null, 2)); + for (const error3 of errors) { + logger.startGroup(`Error details: ${error3.stack}`); + logger.info(JSON.stringify(error3, null, 2)); logger.endGroup(); } const sarifErrors = errors.map((e) => `- ${e.stack}`); @@ -93693,10 +93706,10 @@ function shouldConsiderConfigurationError(processingErrors) { } function shouldConsiderInvalidRequest(processingErrors) { return processingErrors.every( - (error4) => error4.startsWith("rejecting SARIF") || error4.startsWith("an invalid URI was provided as a SARIF location") || error4.startsWith("locationFromSarifResult: expected artifact location") || error4.startsWith( + (error3) => error3.startsWith("rejecting SARIF") || error3.startsWith("an invalid URI was provided as a SARIF location") || error3.startsWith("locationFromSarifResult: expected artifact location") || error3.startsWith( "could not convert rules: invalid security severity value, is not a number" ) || /^SARIF URI scheme [^\s]* did not match the checkout URI scheme [^\s]*/.test( - error4 + error3 ) ); } @@ -93812,8 +93825,8 @@ async function postProcessAndUploadSarif(logger, features, uploadKind, checkoutP } // src/analyze-action.ts -async function sendStatusReport2(startedAt, config, stats, error4, trapCacheUploadTime, dbCreationTimings, didUploadTrapCaches, trapCacheCleanup, dependencyCacheResults, logger) { - const status = getActionsStatus(error4, stats?.analyze_failure_language); +async function sendStatusReport2(startedAt, config, stats, error3, trapCacheUploadTime, dbCreationTimings, didUploadTrapCaches, trapCacheCleanup, dependencyCacheResults, logger) { + const status = getActionsStatus(error3, stats?.analyze_failure_language); const statusReportBase = await createStatusReportBase( "finish" /* Analyze */, status, @@ -93821,8 +93834,8 @@ async function sendStatusReport2(startedAt, config, stats, error4, trapCacheUplo config, await checkDiskUsage(logger), logger, - error4?.message, - error4?.stack + error3?.message, + error3?.stack ); if (statusReportBase !== void 0) { const report = { @@ -94067,8 +94080,15 @@ async function run() { } else { logger.info("Not uploading results"); } - await uploadOverlayBaseDatabaseToCache(codeql, config, logger); - await uploadDatabases(repositoryNwo, codeql, config, apiDetails, logger); + await cleanupAndUploadOverlayBaseDatabaseToCache(codeql, config, logger); + await cleanupAndUploadDatabases( + repositoryNwo, + codeql, + config, + apiDetails, + features, + logger + ); const trapCacheUploadStartTime = import_perf_hooks3.performance.now(); didUploadTrapCaches = await uploadTrapCaches(codeql, config, logger); trapCacheUploadTime = import_perf_hooks3.performance.now() - trapCacheUploadStartTime; @@ -94101,15 +94121,15 @@ async function run() { } core14.exportVariable("CODEQL_ACTION_ANALYZE_DID_COMPLETE_SUCCESSFULLY" /* ANALYZE_DID_COMPLETE_SUCCESSFULLY */, "true"); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); if (getOptionalInput("expect-error") !== "true" || hasBadExpectErrorInput()) { - core14.setFailed(error4.message); + core14.setFailed(error3.message); } await sendStatusReport2( startedAt, config, - error4 instanceof CodeQLAnalysisError ? error4.queriesStatusReport : void 0, - error4 instanceof CodeQLAnalysisError ? error4.error : error4, + error3 instanceof CodeQLAnalysisError ? error3.queriesStatusReport : void 0, + error3 instanceof CodeQLAnalysisError ? error3.error : error3, trapCacheUploadTime, dbCreationTimings, didUploadTrapCaches, @@ -94167,8 +94187,8 @@ var runPromise = run(); async function runWrapper() { try { await runPromise; - } catch (error4) { - core14.setFailed(`analyze action failed: ${getErrorMessage(error4)}`); + } catch (error3) { + core14.setFailed(`analyze action failed: ${getErrorMessage(error3)}`); } await checkForTimeout(); } @@ -94186,7 +94206,7 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) long/index.js: (** diff --git a/lib/autobuild-action.js b/lib/autobuild-action.js index 48485a850..e9f13367e 100644 --- a/lib/autobuild-action.js +++ b/lib/autobuild-action.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug5("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay === "number" && delay > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve5(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug5; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning8(message, properties = {}) { + exports2.error = error3; + function warning9(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning8; + exports2.warning = warning9; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { @@ -27662,23 +27662,21 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -27686,10 +27684,10 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", @@ -27699,9 +27697,9 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27723,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -27911,8 +27910,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27921,8 @@ var require_light = __commonJS({ return (await Promise.all(promises2)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28034,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28071,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run2, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28089,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } } doExpire(clearGlobalState, run2, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run2, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run2, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28116,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28395,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve5, returned, task; + var args, cb, error3, reject, resolve5, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve5, reject } = this._queue.shift()); @@ -28407,9 +28406,9 @@ var require_light = __commonJS({ return resolve5(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28542,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28876,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29179,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29212,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29237,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -33935,7 +33934,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core14 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob = __importStar4(require_glob()); var io5 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -34019,7 +34018,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core14.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -36085,8 +36084,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36320,9 +36319,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37653,14 +37652,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37670,7 +37669,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37678,8 +37677,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -37899,7 +37898,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39109,14 +39108,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39788,11 +39787,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39807,8 +39806,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40302,8 +40301,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40537,9 +40536,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41039,8 +41038,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41274,9 +41273,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42339,9 +42338,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42389,13 +42388,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42415,21 +42414,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42594,8 +42593,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43001,16 +43000,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43554,10 +43553,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45636,12 +45635,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -45906,16 +45905,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46040,7 +46039,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46206,7 +46205,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46447,9 +46446,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47157,7 +47156,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48220,26 +48219,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48280,12 +48279,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48293,13 +48292,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48308,7 +48307,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -62914,8 +62913,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -62949,10 +62948,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64536,8 +64535,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64552,9 +64551,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve5, reject) => { this.emitter.on("finish", resolve5); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65655,8 +65654,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65740,7 +65739,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -68892,7 +68891,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70259,9 +70258,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70375,12 +70374,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70414,13 +70413,13 @@ var require_requestUtils = __commonJS({ delay, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71236,8 +71235,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73083,8 +73082,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74655,12 +74654,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74680,12 +74679,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75149,8 +75148,8 @@ var require_test_transport = __commonJS({ } try { yield delay(this.responseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75161,8 +75160,8 @@ var require_test_transport = __commonJS({ stream.notifyMessage(msg); try { yield delay(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } } @@ -76225,8 +76224,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76322,8 +76321,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76354,18 +76353,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76633,8 +76632,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76835,22 +76834,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76905,15 +76904,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -76921,8 +76920,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76984,10 +76983,10 @@ var require_cache3 = __commonJS({ } core14.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core14.info(`Failed to save: ${typedError.message}`); } else { @@ -77000,8 +76999,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77046,8 +77045,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core14.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core14.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core14.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77066,10 +77065,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core14.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77084,8 +77083,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79810,7 @@ var require_debug2 = __commonJS({ if (!debug5) { try { debug5 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug5 !== "function") { debug5 = function() { @@ -79844,8 +79843,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79918,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80087,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80289,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -80333,7 +80332,6 @@ var io2 = __toESM(require_io2()); var fsPromises = __toESM(require("fs/promises")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js @@ -80351,14 +80349,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs7.lstat(itemPath, { bigint: true }) : await fs7.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs7.lstat(itemPath, { bigint: true }) : await fs7.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs7.readdir(itemPath) : await fs7.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs7.readdir(itemPath) : await fs7.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -80369,13 +80367,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -81255,6 +81253,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -81374,7 +81384,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -81414,16 +81424,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -82993,9 +82994,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -83123,11 +83124,11 @@ function getTestingEnvironment() { } return testingEnvironment; } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } async function checkDiskUsage(logger) { try { @@ -83150,9 +83151,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -83164,7 +83165,7 @@ function checkActionVersion(version, githubVersion) { semver.coerce(githubVersion.version) ?? "0.0.0", ">=3.20" )) { - core3.error( + core3.warning( "CodeQL Action v3 will be deprecated in December 2026. Please update all occurrences of the CodeQL Action in your workflow files to v4. For more information, see https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/" ); core3.exportVariable("CODEQL_ACTION_DID_LOG_VERSION_DEPRECATION" /* LOG_VERSION_DEPRECATION */, "true"); @@ -83459,19 +83460,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -83487,9 +83488,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -83700,8 +83701,8 @@ var path3 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/overlay-database-utils.ts var fs2 = __toESM(require("fs")); @@ -83730,13 +83731,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -83886,7 +83887,7 @@ function getActionsLogger() { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { @@ -84015,6 +84016,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -84126,21 +84132,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, @@ -84418,6 +84424,8 @@ var GitHubFeatureFlags = class { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, @@ -85188,9 +85196,9 @@ function isFirstPartyAnalysis(actionName) { } return process.env["CODEQL_ACTION_INIT_HAS_RUN" /* INIT_ACTION_HAS_RUN */] === "true"; } -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -85427,9 +85435,9 @@ async function run() { } await endTracingForCluster(codeql, config, logger); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); core13.setFailed( - `We were unable to automatically build your code. Please replace the call to the autobuild action with your custom build steps. ${error4.message}` + `We were unable to automatically build your code. Please replace the call to the autobuild action with your custom build steps. ${error3.message}` ); await sendCompletedStatusReport( config, @@ -85437,7 +85445,7 @@ async function run() { startedAt, languages ?? [], currentLanguage, - error4 + error3 ); return; } @@ -85447,8 +85455,8 @@ async function run() { async function runWrapper() { try { await run(); - } catch (error4) { - core13.setFailed(`autobuild action failed. ${getErrorMessage(error4)}`); + } catch (error3) { + core13.setFailed(`autobuild action failed. ${getErrorMessage(error3)}`); } } void runWrapper(); @@ -85461,5 +85469,5 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/defaults.json b/lib/defaults.json index 1fa392f54..835b6a33b 100644 --- a/lib/defaults.json +++ b/lib/defaults.json @@ -1,6 +1,6 @@ { - "bundleVersion": "codeql-bundle-v2.23.3", - "cliVersion": "2.23.3", - "priorBundleVersion": "codeql-bundle-v2.23.2", - "priorCliVersion": "2.23.2" + "bundleVersion": "codeql-bundle-v2.23.6", + "cliVersion": "2.23.6", + "priorBundleVersion": "codeql-bundle-v2.23.5", + "priorCliVersion": "2.23.5" } diff --git a/lib/init-action-post.js b/lib/init-action-post.js index b15b16f37..2bcddfbfa 100644 --- a/lib/init-action-post.js +++ b/lib/init-action-post.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug5("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve8(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug5; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning11(message, properties = {}) { + exports2.error = error3; + function warning12(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning11; + exports2.warning = warning12; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url2 = ""; return { value: { @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { @@ -27662,23 +27662,21 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -27686,10 +27684,10 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", @@ -27699,9 +27697,9 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27723,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -27911,8 +27910,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27921,8 @@ var require_light = __commonJS({ return (await Promise.all(promises5)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28034,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28071,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run2, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28089,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } } doExpire(clearGlobalState, run2, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run2, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run2, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28116,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28395,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve8, returned, task; + var args, cb, error3, reject, resolve8, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve8, reject } = this._queue.shift()); @@ -28407,9 +28406,9 @@ var require_light = __commonJS({ return resolve8(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28542,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28876,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29179,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29212,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29237,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -33935,7 +33934,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core18 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob2 = __importStar4(require_glob()); var io7 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -34019,7 +34018,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core18.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -36085,8 +36084,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36320,9 +36319,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37653,14 +37652,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37670,7 +37669,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37678,8 +37677,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -37899,7 +37898,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39109,14 +39108,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39788,11 +39787,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39807,8 +39806,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40302,8 +40301,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40537,9 +40536,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41039,8 +41038,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41274,9 +41273,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42339,9 +42338,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42389,13 +42388,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42415,21 +42414,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42594,8 +42593,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43001,16 +43000,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43554,10 +43553,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45636,12 +45635,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -45906,16 +45905,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46040,7 +46039,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46206,7 +46205,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46447,9 +46446,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47157,7 +47156,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48220,26 +48219,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48280,12 +48279,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48293,13 +48292,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48308,7 +48307,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -62914,8 +62913,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -62949,10 +62948,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64536,8 +64535,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64552,9 +64551,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve8, reject) => { this.emitter.on("finish", resolve8); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65655,8 +65654,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65740,7 +65739,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -68892,7 +68891,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70259,9 +70258,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core18.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core18.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70375,12 +70374,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70414,13 +70413,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71236,8 +71235,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73083,8 +73082,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74655,12 +74654,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74680,12 +74679,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75149,8 +75148,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75161,8 +75160,8 @@ var require_test_transport = __commonJS({ stream2.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } } @@ -76225,8 +76224,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url2} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url2} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76322,8 +76321,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url2, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76354,18 +76353,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76633,8 +76632,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76835,22 +76834,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core18.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core18.error(`Failed to restore: ${error4.message}`); + core18.error(`Failed to restore: ${error3.message}`); } else { - core18.warning(`Failed to restore: ${error4.message}`); + core18.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core18.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core18.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76905,15 +76904,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core18.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core18.error(`Failed to restore: ${error4.message}`); + core18.error(`Failed to restore: ${error3.message}`); } else { - core18.warning(`Failed to restore: ${error4.message}`); + core18.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -76921,8 +76920,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core18.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core18.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76984,10 +76983,10 @@ var require_cache3 = __commonJS({ } core18.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core18.info(`Failed to save: ${typedError.message}`); } else { @@ -77000,8 +76999,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core18.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core18.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77046,8 +77045,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core18.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core18.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError2(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core18.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77066,10 +77065,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core18.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77084,8 +77083,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core18.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core18.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79810,7 @@ var require_debug2 = __commonJS({ if (!debug5) { try { debug5 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug5 !== "function") { debug5 = function() { @@ -79844,8 +79843,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79918,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80087,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80289,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -83638,8 +83637,8 @@ var require_util11 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url2} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url2} ${error3 instanceof Error ? error3.message : String(error3)}`); } } function maskSecretUrls(body) { @@ -83732,8 +83731,8 @@ var require_artifact_twirp_client2 = __commonJS({ return this.httpClient.post(url2, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -83764,18 +83763,18 @@ var require_artifact_twirp_client2 = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -84046,11 +84045,11 @@ var require_blob_upload = __commonJS({ blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options), chunkTimer((0, config_1.getUploadChunkTimeout)()) ]); - } catch (error4) { - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + } catch (error3) { + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } - throw error4; + throw error3; } finally { abortController.abort(); } @@ -85071,9 +85070,9 @@ var require_async = __commonJS({ invokeCallback(callback, err && (err instanceof Error || err.message) ? err : new Error(err)); }); } - function invokeCallback(callback, error4, value) { + function invokeCallback(callback, error3, value) { try { - callback(error4, value); + callback(error3, value); } catch (err) { setImmediate$1((e) => { throw e; @@ -86379,10 +86378,10 @@ var require_async = __commonJS({ function reflect(fn) { var _fn = wrapAsync(fn); return initialParams(function reflectOn(args, reflectCallback) { - args.push((error4, ...cbArgs) => { + args.push((error3, ...cbArgs) => { let retVal = {}; - if (error4) { - retVal.error = error4; + if (error3) { + retVal.error = error3; } if (cbArgs.length > 0) { var value = cbArgs; @@ -86538,13 +86537,13 @@ var require_async = __commonJS({ var timer; function timeoutCallback() { var name = asyncFn.name || "anonymous"; - var error4 = new Error('Callback function "' + name + '" timed out.'); - error4.code = "ETIMEDOUT"; + var error3 = new Error('Callback function "' + name + '" timed out.'); + error3.code = "ETIMEDOUT"; if (info7) { - error4.info = info7; + error3.info = info7; } timedOut = true; - callback(error4); + callback(error3); } args.push((...cbArgs) => { if (!timedOut) { @@ -86587,7 +86586,7 @@ var require_async = __commonJS({ return callback[PROMISE_SYMBOL]; } function tryEach(tasks, callback) { - var error4 = null; + var error3 = null; var result; return eachSeries$1(tasks, (task, taskCb) => { wrapAsync(task)((err, ...args) => { @@ -86597,10 +86596,10 @@ var require_async = __commonJS({ } else { result = args; } - error4 = err; + error3 = err; taskCb(err ? null : {}); }); - }, () => callback(error4, result)); + }, () => callback(error3, result)); } var tryEach$1 = awaitify(tryEach); function unmemoize(fn) { @@ -93080,19 +93079,19 @@ var require_from = __commonJS({ next(); } }; - readable._destroy = function(error4, cb) { + readable._destroy = function(error3, cb) { PromisePrototypeThen( - close(error4), - () => process2.nextTick(cb, error4), + close(error3), + () => process2.nextTick(cb, error3), // nextTick is here in case cb throws - (e) => process2.nextTick(cb, e || error4) + (e) => process2.nextTick(cb, e || error3) ); }; - async function close(error4) { - const hadError = error4 !== void 0 && error4 !== null; + async function close(error3) { + const hadError = error3 !== void 0 && error3 !== null; const hasThrow = typeof iterator.throw === "function"; if (hadError && hasThrow) { - const { value, done } = await iterator.throw(error4); + const { value, done } = await iterator.throw(error3); await value; if (done) { return; @@ -93300,12 +93299,12 @@ var require_readable3 = __commonJS({ this.destroy(err); }; Readable2.prototype[SymbolAsyncDispose] = function() { - let error4; + let error3; if (!this.destroyed) { - error4 = this.readableEnded ? null : new AbortError(); - this.destroy(error4); + error3 = this.readableEnded ? null : new AbortError(); + this.destroy(error3); } - return new Promise2((resolve8, reject) => eos(this, (err) => err && err !== error4 ? reject(err) : resolve8(null))); + return new Promise2((resolve8, reject) => eos(this, (err) => err && err !== error3 ? reject(err) : resolve8(null))); }; Readable2.prototype.push = function(chunk, encoding) { return readableAddChunk(this, chunk, encoding, false); @@ -93858,14 +93857,14 @@ var require_readable3 = __commonJS({ } } stream2.on("readable", next); - let error4; + let error3; const cleanup = eos( stream2, { writable: false }, (err) => { - error4 = err ? aggregateTwoErrors(error4, err) : null; + error3 = err ? aggregateTwoErrors(error3, err) : null; callback(); callback = nop; } @@ -93875,19 +93874,19 @@ var require_readable3 = __commonJS({ const chunk = stream2.destroyed ? null : stream2.read(); if (chunk !== null) { yield chunk; - } else if (error4) { - throw error4; - } else if (error4 === null) { + } else if (error3) { + throw error3; + } else if (error3 === null) { return; } else { await new Promise2(next); } } } catch (err) { - error4 = aggregateTwoErrors(error4, err); - throw error4; + error3 = aggregateTwoErrors(error3, err); + throw error3; } finally { - if ((error4 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error4 === void 0 || stream2._readableState.autoDestroy)) { + if ((error3 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error3 === void 0 || stream2._readableState.autoDestroy)) { destroyImpl.destroyer(stream2, null); } else { stream2.off("readable", next); @@ -95380,11 +95379,11 @@ var require_pipeline3 = __commonJS({ yield* Readable2.prototype[SymbolAsyncIterator].call(val2); } async function pumpToNode(iterable, writable, finish, { end }) { - let error4; + let error3; let onresolve = null; const resume = (err) => { if (err) { - error4 = err; + error3 = err; } if (onresolve) { const callback = onresolve; @@ -95393,12 +95392,12 @@ var require_pipeline3 = __commonJS({ } }; const wait = () => new Promise2((resolve8, reject) => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { onresolve = () => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve8(); } @@ -95428,7 +95427,7 @@ var require_pipeline3 = __commonJS({ } finish(); } catch (err) { - finish(error4 !== err ? aggregateTwoErrors(error4, err) : err); + finish(error3 !== err ? aggregateTwoErrors(error3, err) : err); } finally { cleanup(); writable.off("drain", resume); @@ -95482,7 +95481,7 @@ var require_pipeline3 = __commonJS({ if (outerSignal) { disposable = addAbortListener(outerSignal, abort); } - let error4; + let error3; let value; const destroys = []; let finishCount = 0; @@ -95491,23 +95490,23 @@ var require_pipeline3 = __commonJS({ } function finishImpl(err, final) { var _disposable; - if (err && (!error4 || error4.code === "ERR_STREAM_PREMATURE_CLOSE")) { - error4 = err; + if (err && (!error3 || error3.code === "ERR_STREAM_PREMATURE_CLOSE")) { + error3 = err; } - if (!error4 && !final) { + if (!error3 && !final) { return; } while (destroys.length) { - destroys.shift()(error4); + destroys.shift()(error3); } ; (_disposable = disposable) === null || _disposable === void 0 ? void 0 : _disposable[SymbolDispose](); ac.abort(); if (final) { - if (!error4) { + if (!error3) { lastStreamCleanup.forEach((fn) => fn()); } - process2.nextTick(callback, error4, value); + process2.nextTick(callback, error3, value); } } let ret; @@ -97357,52 +97356,128 @@ var require_isPlainObject = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/brace-expansion/index.js -var require_brace_expansion3 = __commonJS({ - "node_modules/archiver-utils/node_modules/brace-expansion/index.js"(exports2, module2) { - var balanced = require_balanced_match(); - module2.exports = expandTop; +// node_modules/@isaacs/balanced-match/dist/commonjs/index.js +var require_commonjs13 = __commonJS({ + "node_modules/@isaacs/balanced-match/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.range = exports2.balanced = void 0; + var balanced = (a, b, str2) => { + const ma = a instanceof RegExp ? maybeMatch(a, str2) : a; + const mb = b instanceof RegExp ? maybeMatch(b, str2) : b; + const r = ma !== null && mb != null && (0, exports2.range)(ma, mb, str2); + return r && { + start: r[0], + end: r[1], + pre: str2.slice(0, r[0]), + body: str2.slice(r[0] + ma.length, r[1]), + post: str2.slice(r[1] + mb.length) + }; + }; + exports2.balanced = balanced; + var maybeMatch = (reg, str2) => { + const m = str2.match(reg); + return m ? m[0] : null; + }; + var range = (a, b, str2) => { + let begs, beg, left, right = void 0, result; + let ai = str2.indexOf(a); + let bi = str2.indexOf(b, ai + 1); + let i = ai; + if (ai >= 0 && bi > 0) { + if (a === b) { + return [ai, bi]; + } + begs = []; + left = str2.length; + while (i >= 0 && !result) { + if (i === ai) { + begs.push(i); + ai = str2.indexOf(a, i + 1); + } else if (begs.length === 1) { + const r = begs.pop(); + if (r !== void 0) + result = [r, bi]; + } else { + beg = begs.pop(); + if (beg !== void 0 && beg < left) { + left = beg; + right = bi; + } + bi = str2.indexOf(b, i + 1); + } + i = ai < bi && ai >= 0 ? ai : bi; + } + if (begs.length && right !== void 0) { + result = [left, right]; + } + } + return result; + }; + exports2.range = range; + } +}); + +// node_modules/@isaacs/brace-expansion/dist/commonjs/index.js +var require_commonjs14 = __commonJS({ + "node_modules/@isaacs/brace-expansion/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.expand = expand; + var balanced_match_1 = require_commonjs13(); var escSlash = "\0SLASH" + Math.random() + "\0"; var escOpen = "\0OPEN" + Math.random() + "\0"; var escClose = "\0CLOSE" + Math.random() + "\0"; var escComma = "\0COMMA" + Math.random() + "\0"; var escPeriod = "\0PERIOD" + Math.random() + "\0"; + var escSlashPattern = new RegExp(escSlash, "g"); + var escOpenPattern = new RegExp(escOpen, "g"); + var escClosePattern = new RegExp(escClose, "g"); + var escCommaPattern = new RegExp(escComma, "g"); + var escPeriodPattern = new RegExp(escPeriod, "g"); + var slashPattern = /\\\\/g; + var openPattern = /\\{/g; + var closePattern = /\\}/g; + var commaPattern = /\\,/g; + var periodPattern = /\\./g; function numeric(str2) { - return parseInt(str2, 10) == str2 ? parseInt(str2, 10) : str2.charCodeAt(0); + return !isNaN(str2) ? parseInt(str2, 10) : str2.charCodeAt(0); } function escapeBraces(str2) { - return str2.split("\\\\").join(escSlash).split("\\{").join(escOpen).split("\\}").join(escClose).split("\\,").join(escComma).split("\\.").join(escPeriod); + return str2.replace(slashPattern, escSlash).replace(openPattern, escOpen).replace(closePattern, escClose).replace(commaPattern, escComma).replace(periodPattern, escPeriod); } function unescapeBraces(str2) { - return str2.split(escSlash).join("\\").split(escOpen).join("{").split(escClose).join("}").split(escComma).join(",").split(escPeriod).join("."); + return str2.replace(escSlashPattern, "\\").replace(escOpenPattern, "{").replace(escClosePattern, "}").replace(escCommaPattern, ",").replace(escPeriodPattern, "."); } function parseCommaParts(str2) { - if (!str2) + if (!str2) { return [""]; - var parts = []; - var m = balanced("{", "}", str2); - if (!m) + } + const parts = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) { return str2.split(","); - var pre = m.pre; - var body = m.body; - var post = m.post; - var p = pre.split(","); + } + const { pre, body, post } = m; + const p = pre.split(","); p[p.length - 1] += "{" + body + "}"; - var postParts = parseCommaParts(post); + const postParts = parseCommaParts(post); if (post.length) { + ; p[p.length - 1] += postParts.shift(); p.push.apply(p, postParts); } parts.push.apply(parts, p); return parts; } - function expandTop(str2) { - if (!str2) + function expand(str2) { + if (!str2) { return []; - if (str2.substr(0, 2) === "{}") { - str2 = "\\{\\}" + str2.substr(2); } - return expand(escapeBraces(str2), true).map(unescapeBraces); + if (str2.slice(0, 2) === "{}") { + str2 = "\\{\\}" + str2.slice(2); + } + return expand_(escapeBraces(str2), true).map(unescapeBraces); } function embrace(str2) { return "{" + str2 + "}"; @@ -97416,73 +97491,74 @@ var require_brace_expansion3 = __commonJS({ function gte5(i, y) { return i >= y; } - function expand(str2, isTop) { - var expansions = []; - var m = balanced("{", "}", str2); - if (!m) return [str2]; - var pre = m.pre; - var post = m.post.length ? expand(m.post, false) : [""]; + function expand_(str2, isTop) { + const expansions = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) + return [str2]; + const pre = m.pre; + const post = m.post.length ? expand_(m.post, false) : [""]; if (/\$$/.test(m.pre)) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + "{" + m.body + "}" + post[k]; + for (let k = 0; k < post.length; k++) { + const expansion = pre + "{" + m.body + "}" + post[k]; expansions.push(expansion); } } else { - var isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); - var isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); - var isSequence = isNumericSequence || isAlphaSequence; - var isOptions = m.body.indexOf(",") >= 0; + const isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); + const isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); + const isSequence = isNumericSequence || isAlphaSequence; + const isOptions = m.body.indexOf(",") >= 0; if (!isSequence && !isOptions) { if (m.post.match(/,(?!,).*\}/)) { str2 = m.pre + "{" + m.body + escClose + m.post; - return expand(str2); + return expand_(str2); } return [str2]; } - var n; + let n; if (isSequence) { n = m.body.split(/\.\./); } else { n = parseCommaParts(m.body); - if (n.length === 1) { - n = expand(n[0], false).map(embrace); + if (n.length === 1 && n[0] !== void 0) { + n = expand_(n[0], false).map(embrace); if (n.length === 1) { - return post.map(function(p) { - return m.pre + n[0] + p; - }); + return post.map((p) => m.pre + n[0] + p); } } } - var N; - if (isSequence) { - var x = numeric(n[0]); - var y = numeric(n[1]); - var width = Math.max(n[0].length, n[1].length); - var incr = n.length == 3 ? Math.abs(numeric(n[2])) : 1; - var test = lte; - var reverse = y < x; + let N; + if (isSequence && n[0] !== void 0 && n[1] !== void 0) { + const x = numeric(n[0]); + const y = numeric(n[1]); + const width = Math.max(n[0].length, n[1].length); + let incr = n.length === 3 && n[2] !== void 0 ? Math.abs(numeric(n[2])) : 1; + let test = lte; + const reverse = y < x; if (reverse) { incr *= -1; test = gte5; } - var pad = n.some(isPadded); + const pad = n.some(isPadded); N = []; - for (var i = x; test(i, y); i += incr) { - var c; + for (let i = x; test(i, y); i += incr) { + let c; if (isAlphaSequence) { c = String.fromCharCode(i); - if (c === "\\") + if (c === "\\") { c = ""; + } } else { c = String(i); if (pad) { - var need = width - c.length; + const need = width - c.length; if (need > 0) { - var z = new Array(need + 1).join("0"); - if (i < 0) + const z = new Array(need + 1).join("0"); + if (i < 0) { c = "-" + z + c.slice(1); - else + } else { c = z + c; + } } } } @@ -97490,15 +97566,16 @@ var require_brace_expansion3 = __commonJS({ } } else { N = []; - for (var j = 0; j < n.length; j++) { - N.push.apply(N, expand(n[j], false)); + for (let j = 0; j < n.length; j++) { + N.push.apply(N, expand_(n[j], false)); } } - for (var j = 0; j < N.length; j++) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + N[j] + post[k]; - if (!isTop || isSequence || expansion) + for (let j = 0; j < N.length; j++) { + for (let k = 0; k < post.length; k++) { + const expansion = pre + N[j] + post[k]; + if (!isTop || isSequence || expansion) { expansions.push(expansion); + } } } } @@ -97507,9 +97584,9 @@ var require_brace_expansion3 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js var require_assert_valid_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.assertValidPattern = void 0; @@ -97526,9 +97603,9 @@ var require_assert_valid_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js var require_brace_expressions = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.parseClass = void 0; @@ -97643,22 +97720,25 @@ var require_brace_expressions = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js var require_unescape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = void 0; - var unescape = (s, { windowsPathsNoEscape = false } = {}) => { - return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + var unescape = (s, { windowsPathsNoEscape = false, magicalBraces = true } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + } + return windowsPathsNoEscape ? s.replace(/\[([^\/\\{}])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\{}])\]/g, "$1$2").replace(/\\([^\/{}])/g, "$1"); }; exports2.unescape = unescape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js var require_ast = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.AST = void 0; @@ -98014,7 +98094,7 @@ var require_ast = __commonJS({ if (this.#root === this) this.#fillNegs(); if (!this.type) { - const noEmpty = this.isStart() && this.isEnd(); + const noEmpty = this.isStart() && this.isEnd() && !this.#parts.some((s) => typeof s !== "string"); const src = this.#parts.map((p) => { const [re, _2, hasMagic, uflag] = typeof p === "string" ? _AST.#parseGlob(p, this.#hasMagic, noEmpty) : p.toRegExpSource(allowDot); this.#hasMagic = this.#hasMagic || hasMagic; @@ -98124,10 +98204,7 @@ var require_ast = __commonJS({ } } if (c === "*") { - if (noEmpty && glob2 === "*") - re += starNoEmpty; - else - re += star; + re += noEmpty && glob2 === "*" ? starNoEmpty : star; hasMagic = true; continue; } @@ -98145,29 +98222,29 @@ var require_ast = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js var require_escape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.escape = void 0; - var escape = (s, { windowsPathsNoEscape = false } = {}) => { + var escape = (s, { windowsPathsNoEscape = false, magicalBraces = false } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/[?*()[\]{}]/g, "[$&]") : s.replace(/[?*()[\]\\{}]/g, "\\$&"); + } return windowsPathsNoEscape ? s.replace(/[?*()[\]]/g, "[$&]") : s.replace(/[?*()[\]\\]/g, "\\$&"); }; exports2.escape = escape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js -var require_commonjs13 = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js"(exports2) { +// node_modules/glob/node_modules/minimatch/dist/commonjs/index.js +var require_commonjs15 = __commonJS({ + "node_modules/glob/node_modules/minimatch/dist/commonjs/index.js"(exports2) { "use strict"; - var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { - return mod && mod.__esModule ? mod : { "default": mod }; - }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = exports2.escape = exports2.AST = exports2.Minimatch = exports2.match = exports2.makeRe = exports2.braceExpand = exports2.defaults = exports2.filter = exports2.GLOBSTAR = exports2.sep = exports2.minimatch = void 0; - var brace_expansion_1 = __importDefault4(require_brace_expansion3()); + var brace_expansion_1 = require_commonjs14(); var assert_valid_pattern_js_1 = require_assert_valid_pattern(); var ast_js_1 = require_ast(); var escape_js_1 = require_escape(); @@ -98290,7 +98367,7 @@ var require_commonjs13 = __commonJS({ if (options.nobrace || !/\{(?:(?!\{).)*\}/.test(pattern)) { return [pattern]; } - return (0, brace_expansion_1.default)(pattern); + return (0, brace_expansion_1.expand)(pattern); }; exports2.braceExpand = braceExpand; exports2.minimatch.braceExpand = exports2.braceExpand; @@ -98814,16 +98891,27 @@ var require_commonjs13 = __commonJS({ pp[i] = twoStar; } } else if (next === void 0) { - pp[i - 1] = prev + "(?:\\/|" + twoStar + ")?"; + pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + ")?"; } else if (next !== exports2.GLOBSTAR) { pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + "\\/)" + next; pp[i + 1] = exports2.GLOBSTAR; } }); - return pp.filter((p) => p !== exports2.GLOBSTAR).join("/"); + const filtered = pp.filter((p) => p !== exports2.GLOBSTAR); + if (this.partial && filtered.length >= 1) { + const prefixes = []; + for (let i = 1; i <= filtered.length; i++) { + prefixes.push(filtered.slice(0, i).join("/")); + } + return "(?:" + prefixes.join("|") + ")"; + } + return filtered.join("/"); }).join("|"); const [open, close] = set2.length > 1 ? ["(?:", ")"] : ["", ""]; re = "^" + open + re + close + "$"; + if (this.partial) { + re = "^(?:\\/|" + open + re.slice(1, -1) + close + ")$"; + } if (this.negate) re = "^(?!" + re + ").+$"; try { @@ -98910,9 +98998,9 @@ var require_commonjs13 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js -var require_commonjs14 = __commonJS({ - "node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js"(exports2) { +// node_modules/lru-cache/dist/commonjs/index.js +var require_commonjs16 = __commonJS({ + "node_modules/lru-cache/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.LRUCache = void 0; @@ -99001,6 +99089,7 @@ var require_commonjs14 = __commonJS({ #max; #maxSize; #dispose; + #onInsert; #disposeAfter; #fetchMethod; #memoMethod; @@ -99082,6 +99171,7 @@ var require_commonjs14 = __commonJS({ #hasDispose; #hasFetchMethod; #hasDisposeAfter; + #hasOnInsert; /** * Do not call this method unless you need to inspect the * inner workings of the cache. If anything returned by this @@ -99158,6 +99248,12 @@ var require_commonjs14 = __commonJS({ get dispose() { return this.#dispose; } + /** + * {@link LRUCache.OptionsBase.onInsert} (read-only) + */ + get onInsert() { + return this.#onInsert; + } /** * {@link LRUCache.OptionsBase.disposeAfter} (read-only) */ @@ -99165,7 +99261,7 @@ var require_commonjs14 = __commonJS({ return this.#disposeAfter; } constructor(options) { - const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; + const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, onInsert, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; if (max !== 0 && !isPosInt(max)) { throw new TypeError("max option must be a nonnegative integer"); } @@ -99207,6 +99303,9 @@ var require_commonjs14 = __commonJS({ if (typeof dispose === "function") { this.#dispose = dispose; } + if (typeof onInsert === "function") { + this.#onInsert = onInsert; + } if (typeof disposeAfter === "function") { this.#disposeAfter = disposeAfter; this.#disposed = []; @@ -99215,6 +99314,7 @@ var require_commonjs14 = __commonJS({ this.#disposed = void 0; } this.#hasDispose = !!this.#dispose; + this.#hasOnInsert = !!this.#onInsert; this.#hasDisposeAfter = !!this.#disposeAfter; this.noDisposeOnSet = !!noDisposeOnSet; this.noUpdateTTL = !!noUpdateTTL; @@ -99617,7 +99717,7 @@ var require_commonjs14 = __commonJS({ } /** * Return an array of [key, {@link LRUCache.Entry}] tuples which can be - * passed to {@link LRLUCache#load}. + * passed to {@link LRUCache#load}. * * The `start` fields are calculated relative to a portable `Date.now()` * timestamp, even if `performance.now()` is available. @@ -99728,6 +99828,9 @@ var require_commonjs14 = __commonJS({ if (status) status.set = "add"; noUpdateTTL = false; + if (this.#hasOnInsert) { + this.#onInsert?.(v, k, "add"); + } } else { this.#moveToTail(index); const oldVal = this.#valList[index]; @@ -99763,6 +99866,9 @@ var require_commonjs14 = __commonJS({ } else if (status) { status.set = "update"; } + if (this.#hasOnInsert) { + this.onInsert?.(v, k, v === oldVal ? "update" : "replace"); + } } if (ttl !== 0 && !this.#ttls) { this.#initializeTTLTracking(); @@ -100288,7 +100394,7 @@ var require_commonjs14 = __commonJS({ }); // node_modules/minipass/dist/commonjs/index.js -var require_commonjs15 = __commonJS({ +var require_commonjs17 = __commonJS({ "node_modules/minipass/dist/commonjs/index.js"(exports2) { "use strict"; var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { @@ -101179,9 +101285,9 @@ var require_commonjs15 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js -var require_commonjs16 = __commonJS({ - "node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js"(exports2) { +// node_modules/path-scurry/dist/commonjs/index.js +var require_commonjs18 = __commonJS({ + "node_modules/path-scurry/dist/commonjs/index.js"(exports2) { "use strict"; var __createBinding4 = exports2 && exports2.__createBinding || (Object.create ? (function(o, m, k, k2) { if (k2 === void 0) k2 = k; @@ -101212,14 +101318,14 @@ var require_commonjs16 = __commonJS({ }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.PathScurry = exports2.Path = exports2.PathScurryDarwin = exports2.PathScurryPosix = exports2.PathScurryWin32 = exports2.PathScurryBase = exports2.PathPosix = exports2.PathWin32 = exports2.PathBase = exports2.ChildrenCache = exports2.ResolveCache = void 0; - var lru_cache_1 = require_commonjs14(); + var lru_cache_1 = require_commonjs16(); var node_path_1 = require("node:path"); var node_url_1 = require("node:url"); var fs_1 = require("fs"); var actualFS = __importStar4(require("node:fs")); var realpathSync = fs_1.realpathSync.native; var promises_1 = require("node:fs/promises"); - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var defaultFS = { lstatSync: fs_1.lstatSync, readdir: fs_1.readdir, @@ -101434,6 +101540,8 @@ var require_commonjs16 = __commonJS({ /** * Deprecated alias for Dirent['parentPath'] Somewhat counterintuitively, * this property refers to the *parent* path, not the path object itself. + * + * @deprecated */ get path() { return this.parentPath; @@ -102953,13 +103061,13 @@ var require_commonjs16 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js +// node_modules/glob/dist/commonjs/pattern.js var require_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js"(exports2) { + "node_modules/glob/dist/commonjs/pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Pattern = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var isPatternList = (pl) => pl.length >= 1; var isGlobList = (gl) => gl.length >= 1; var Pattern = class _Pattern { @@ -103127,13 +103235,13 @@ var require_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js +// node_modules/glob/dist/commonjs/ignore.js var require_ignore = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js"(exports2) { + "node_modules/glob/dist/commonjs/ignore.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Ignore = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var pattern_js_1 = require_pattern(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; var Ignore = class { @@ -103224,13 +103332,13 @@ var require_ignore = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js +// node_modules/glob/dist/commonjs/processor.js var require_processor = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js"(exports2) { + "node_modules/glob/dist/commonjs/processor.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Processor = exports2.SubWalks = exports2.MatchRecord = exports2.HasWalkedCache = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var HasWalkedCache = class _HasWalkedCache { store; constructor(store = /* @__PURE__ */ new Map()) { @@ -103457,13 +103565,13 @@ var require_processor = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js +// node_modules/glob/dist/commonjs/walker.js var require_walker = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js"(exports2) { + "node_modules/glob/dist/commonjs/walker.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.GlobStream = exports2.GlobWalker = exports2.GlobUtil = void 0; - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var ignore_js_1 = require_ignore(); var processor_js_1 = require_processor(); var makeIgnore = (ignore, opts) => typeof ignore === "string" ? new ignore_js_1.Ignore([ignore], opts) : Array.isArray(ignore) ? new ignore_js_1.Ignore(ignore, opts) : ignore; @@ -103797,15 +103905,15 @@ var require_walker = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js +// node_modules/glob/dist/commonjs/glob.js var require_glob2 = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js"(exports2) { + "node_modules/glob/dist/commonjs/glob.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Glob = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var node_url_1 = require("node:url"); - var path_scurry_1 = require_commonjs16(); + var path_scurry_1 = require_commonjs18(); var pattern_js_1 = require_pattern(); var walker_js_1 = require_walker(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; @@ -104010,13 +104118,13 @@ var require_glob2 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js +// node_modules/glob/dist/commonjs/has-magic.js var require_has_magic = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js"(exports2) { + "node_modules/glob/dist/commonjs/has-magic.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.hasMagic = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var hasMagic = (pattern, options = {}) => { if (!Array.isArray(pattern)) { pattern = [pattern]; @@ -104031,9 +104139,9 @@ var require_has_magic = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js -var require_commonjs17 = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js"(exports2) { +// node_modules/glob/dist/commonjs/index.js +var require_commonjs19 = __commonJS({ + "node_modules/glob/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.glob = exports2.sync = exports2.iterate = exports2.iterateSync = exports2.stream = exports2.streamSync = exports2.Ignore = exports2.hasMagic = exports2.Glob = exports2.unescape = exports2.escape = void 0; @@ -104042,10 +104150,10 @@ var require_commonjs17 = __commonJS({ exports2.globSync = globSync; exports2.globIterateSync = globIterateSync; exports2.globIterate = globIterate; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var glob_js_1 = require_glob2(); var has_magic_js_1 = require_has_magic(); - var minimatch_2 = require_commonjs13(); + var minimatch_2 = require_commonjs15(); Object.defineProperty(exports2, "escape", { enumerable: true, get: function() { return minimatch_2.escape; } }); @@ -104122,7 +104230,7 @@ var require_file3 = __commonJS({ var difference = require_difference(); var union = require_union(); var isPlainObject = require_isPlainObject(); - var glob2 = require_commonjs17(); + var glob2 = require_commonjs19(); var file = module2.exports = {}; var pathSeparatorRe = /[\/\\]/g; var processPatterns = function(patterns, fn) { @@ -105849,18 +105957,18 @@ var require_zip_archive_output_stream = __commonJS({ ZipArchiveOutputStream.prototype._smartStream = function(ae, callback) { var deflate = ae.getMethod() === constants.METHOD_DEFLATED; var process2 = deflate ? new DeflateCRC32Stream(this.options.zlib) : new CRC32Stream(); - var error4 = null; + var error3 = null; function handleStuff() { var digest = process2.digest().readUInt32BE(0); ae.setCrc(digest); ae.setSize(process2.size()); ae.setCompressedSize(process2.size(true)); this._afterAppend(ae); - callback(error4, ae); + callback(error3, ae); } process2.once("end", handleStuff.bind(this)); process2.once("error", function(err) { - error4 = err; + error3 = err; }); process2.pipe(this, { end: false }); return process2; @@ -107226,11 +107334,11 @@ var require_streamx = __commonJS({ } [asyncIterator]() { const stream2 = this; - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("readable", onreadable); this.on("close", onclose); @@ -107262,7 +107370,7 @@ var require_streamx = __commonJS({ } function ondata(data) { if (promiseReject === null) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else if (data === null && (stream2._duplexState & READ_DONE) === 0) promiseReject(STREAM_DESTROYED); else promiseResolve({ value: data, done: data === null }); promiseReject = promiseResolve = null; @@ -107436,7 +107544,7 @@ var require_streamx = __commonJS({ if (all.length < 2) throw new Error("Pipeline requires at least 2 streams"); let src = all[0]; let dest = null; - let error4 = null; + let error3 = null; for (let i = 1; i < all.length; i++) { dest = all[i]; if (isStreamx(src)) { @@ -107451,14 +107559,14 @@ var require_streamx = __commonJS({ let fin = false; const autoDestroy = isStreamx(dest) || !!(dest._writableState && dest._writableState.autoDestroy); dest.on("error", (err) => { - if (error4 === null) error4 = err; + if (error3 === null) error3 = err; }); dest.on("finish", () => { fin = true; - if (!autoDestroy) done(error4); + if (!autoDestroy) done(error3); }); if (autoDestroy) { - dest.on("close", () => done(error4 || (fin ? null : PREMATURE_CLOSE))); + dest.on("close", () => done(error3 || (fin ? null : PREMATURE_CLOSE))); } } return dest; @@ -107471,8 +107579,8 @@ var require_streamx = __commonJS({ } } function onerror(err) { - if (!err || error4) return; - error4 = err; + if (!err || error3) return; + error3 = err; for (const s of all) { s.destroy(err); } @@ -108038,7 +108146,7 @@ var require_extract = __commonJS({ cb(null); } [Symbol.asyncIterator]() { - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; let entryStream = null; @@ -108046,7 +108154,7 @@ var require_extract = __commonJS({ const extract2 = this; this.on("entry", onentry); this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("close", onclose); return { @@ -108070,8 +108178,8 @@ var require_extract = __commonJS({ cb(err); } function onnext(resolve8, reject) { - if (error4) { - return reject(error4); + if (error3) { + return reject(error3); } if (entryStream) { resolve8({ value: entryStream, done: false }); @@ -108097,9 +108205,9 @@ var require_extract = __commonJS({ } } function onclose() { - consumeCallback(error4); + consumeCallback(error3); if (!promiseResolve) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else promiseResolve({ value: void 0, done: true }); promiseResolve = promiseReject = null; } @@ -108995,18 +109103,18 @@ var require_zip2 = __commonJS({ return zipUploadStream; }); } - var zipErrorCallback = (error4) => { + var zipErrorCallback = (error3) => { core18.error("An error has occurred while creating the zip file for upload"); - core18.info(error4); + core18.info(error3); throw new Error("An error has occurred during zip creation for the artifact"); }; - var zipWarningCallback = (error4) => { - if (error4.code === "ENOENT") { + var zipWarningCallback = (error3) => { + if (error3.code === "ENOENT") { core18.warning("ENOENT warning during artifact zip creation. No such file or directory"); - core18.info(error4); + core18.info(error3); } else { - core18.warning(`A non-blocking warning has occurred during artifact zip creation: ${error4.code}`); - core18.info(error4); + core18.warning(`A non-blocking warning has occurred during artifact zip creation: ${error3.code}`); + core18.info(error3); } }; var zipFinishCallback = () => { @@ -110813,8 +110921,8 @@ var require_parser_stream = __commonJS({ this.unzipStream.on("entry", function(entry) { self2.push(entry); }); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(ParserStream, Transform); @@ -110954,8 +111062,8 @@ var require_extract2 = __commonJS({ this.createdDirectories = {}; var self2 = this; this.unzipStream.on("entry", this._processEntry.bind(this)); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(Extract, Transform); @@ -110989,8 +111097,8 @@ var require_extract2 = __commonJS({ self2.unfinishedEntries--; self2._notifyAwaiter(); }); - pipedStream.on("error", function(error4) { - self2.emit("error", error4); + pipedStream.on("error", function(error3) { + self2.emit("error", error3); }); entry.pipe(pipedStream); }; @@ -111125,11 +111233,11 @@ var require_download_artifact = __commonJS({ try { yield promises_1.default.access(path15); return true; - } catch (error4) { - if (error4.code === "ENOENT") { + } catch (error3) { + if (error3.code === "ENOENT") { return false; } else { - throw error4; + throw error3; } } }); @@ -111140,9 +111248,9 @@ var require_download_artifact = __commonJS({ while (retryCount < 5) { try { return yield streamExtractExternal(url2, directory); - } catch (error4) { + } catch (error3) { retryCount++; - core18.debug(`Failed to download artifact after ${retryCount} retries due to ${error4.message}. Retrying in 5 seconds...`); + core18.debug(`Failed to download artifact after ${retryCount} retries due to ${error3.message}. Retrying in 5 seconds...`); yield new Promise((resolve8) => setTimeout(resolve8, 5e3)); } } @@ -111171,10 +111279,10 @@ var require_download_artifact = __commonJS({ const extractStream = passThrough; extractStream.on("data", () => { timer.refresh(); - }).on("error", (error4) => { - core18.debug(`response.message: Artifact download failed: ${error4.message}`); + }).on("error", (error3) => { + core18.debug(`response.message: Artifact download failed: ${error3.message}`); clearTimeout(timer); - reject(error4); + reject(error3); }).pipe(unzip_stream_1.default.Extract({ path: directory })).on("close", () => { clearTimeout(timer); if (hashStream) { @@ -111183,8 +111291,8 @@ var require_download_artifact = __commonJS({ core18.info(`SHA256 digest of downloaded artifact is ${sha256Digest}`); } resolve8({ sha256Digest: `sha256:${sha256Digest}` }); - }).on("error", (error4) => { - reject(error4); + }).on("error", (error3) => { + reject(error3); }); }); }); @@ -111223,8 +111331,8 @@ var require_download_artifact = __commonJS({ core18.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -111266,8 +111374,8 @@ Are you trying to download from a different run? Try specifying a github-token w core18.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -111365,9 +111473,9 @@ var require_dist_node16 = __commonJS({ return request(options).then((response) => { octokit.log.info(`${requestOptions.method} ${path15} - ${response.status} in ${Date.now() - start}ms`); return response; - }).catch((error4) => { - octokit.log.info(`${requestOptions.method} ${path15} - ${error4.status} in ${Date.now() - start}ms`); - throw error4; + }).catch((error3) => { + octokit.log.info(`${requestOptions.method} ${path15} - ${error3.status} in ${Date.now() - start}ms`); + throw error3; }); }); } @@ -111385,22 +111493,22 @@ var require_dist_node17 = __commonJS({ return ex && typeof ex === "object" && "default" in ex ? ex["default"] : ex; } var Bottleneck = _interopDefault(require_light()); - async function errorRequest(octokit, state, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(octokit, state, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } async function wrapRequest(state, request, options) { const limiter = new Bottleneck(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -111422,12 +111530,12 @@ var require_dist_node17 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -111918,13 +112026,13 @@ var require_client2 = __commonJS({ throw new errors_1.GHESNotSupportedError(); } return (0, upload_artifact_1.uploadArtifact)(name, files, rootDirectory, options); - } catch (error4) { - (0, core_1.warning)(`Artifact upload failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Artifact upload failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions is operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -111939,13 +112047,13 @@ If the error persists, please check whether Actions is operating normally at [ht return (0, download_artifact_1.downloadArtifactPublic)(artifactId, repositoryOwner, repositoryName, token, downloadOptions); } return (0, download_artifact_1.downloadArtifactInternal)(artifactId, options); - } catch (error4) { - (0, core_1.warning)(`Download Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Download Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -111960,13 +112068,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, list_artifacts_1.listArtifactsPublic)(workflowRunId, repositoryOwner, repositoryName, token, options === null || options === void 0 ? void 0 : options.latest); } return (0, list_artifacts_1.listArtifactsInternal)(options === null || options === void 0 ? void 0 : options.latest); - } catch (error4) { - (0, core_1.warning)(`Listing Artifacts failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Listing Artifacts failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -111981,13 +112089,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, get_artifact_1.getArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, get_artifact_1.getArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Get Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Get Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -112002,13 +112110,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, delete_artifact_1.deleteArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, delete_artifact_1.deleteArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Delete Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Delete Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -112508,14 +112616,14 @@ var require_tmp = __commonJS({ options.template = _getRelativePathSync("template", options.template, tmpDir); return options; } - function _isEBADF(error4) { - return _isExpectedError(error4, -EBADF, "EBADF"); + function _isEBADF(error3) { + return _isExpectedError(error3, -EBADF, "EBADF"); } - function _isENOENT(error4) { - return _isExpectedError(error4, -ENOENT, "ENOENT"); + function _isENOENT(error3) { + return _isExpectedError(error3, -ENOENT, "ENOENT"); } - function _isExpectedError(error4, errno, code) { - return IS_WIN32 ? error4.code === code : error4.code === code && error4.errno === errno; + function _isExpectedError(error3, errno, code) { + return IS_WIN32 ? error3.code === code : error3.code === code && error3.errno === errno; } function setGracefulCleanup() { _gracefulCleanup = true; @@ -114227,9 +114335,9 @@ var require_upload_gzip = __commonJS({ const size = (yield stat(tempFilePath)).size; resolve8(size); })); - outputStream.on("error", (error4) => { - console.log(error4); - reject(error4); + outputStream.on("error", (error3) => { + console.log(error3); + reject(error3); }); }); }); @@ -114354,9 +114462,9 @@ var require_requestUtils2 = __commonJS({ } isRetryable = (0, utils_1.isRetryableStatusCode)(statusCode); errorMessage = `Artifact service responded with ${statusCode}`; - } catch (error4) { + } catch (error3) { isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { core18.info(`${name} - Error is not retryable`); @@ -114722,9 +114830,9 @@ var require_upload_http_client = __commonJS({ let response; try { response = yield uploadChunkRequest(); - } catch (error4) { + } catch (error3) { core18.info(`An error has been caught http-client index ${httpClientIndex}, retrying the upload`); - console.log(error4); + console.log(error3); if (incrementAndCheckRetryLimit()) { return false; } @@ -114913,8 +115021,8 @@ var require_download_http_client = __commonJS({ } this.statusReporter.incrementProcessedCount(); } - }))).catch((error4) => { - throw new Error(`Unable to download the artifact: ${error4}`); + }))).catch((error3) => { + throw new Error(`Unable to download the artifact: ${error3}`); }).finally(() => { this.statusReporter.stop(); this.downloadHttpManager.disposeAndReplaceAllClients(); @@ -114979,9 +115087,9 @@ var require_download_http_client = __commonJS({ let response; try { response = yield makeDownloadRequest(); - } catch (error4) { + } catch (error3) { core18.info("An error occurred while attempting to download a file"); - console.log(error4); + console.log(error3); yield backOff(); continue; } @@ -114995,7 +115103,7 @@ var require_download_http_client = __commonJS({ } else { forceRetry = true; } - } catch (error4) { + } catch (error3) { forceRetry = true; } } @@ -115021,31 +115129,31 @@ var require_download_http_client = __commonJS({ yield new Promise((resolve8, reject) => { if (isGzip) { const gunzip = zlib3.createGunzip(); - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core18.info(`An error occurred while attempting to read the response stream`); gunzip.close(); destinationStream.close(); - reject(error4); - }).pipe(gunzip).on("error", (error4) => { + reject(error3); + }).pipe(gunzip).on("error", (error3) => { core18.info(`An error occurred while attempting to decompress the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve8(); - }).on("error", (error4) => { + }).on("error", (error3) => { core18.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } else { - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core18.info(`An error occurred while attempting to read the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve8(); - }).on("error", (error4) => { + }).on("error", (error3) => { core18.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } }); @@ -119344,7 +119452,6 @@ var fs = __toESM(require("fs")); var fsPromises = __toESM(require("fs/promises")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js @@ -119362,14 +119469,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs17.lstat(itemPath, { bigint: true }) : await fs17.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs17.lstat(itemPath, { bigint: true }) : await fs17.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs17.readdir(itemPath) : await fs17.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs17.readdir(itemPath) : await fs17.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -119380,13 +119487,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -120266,6 +120373,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -120385,7 +120504,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -120425,16 +120544,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -122005,9 +122115,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -122201,11 +122311,11 @@ function parseMatrixInput(matrixInput) { } return JSON.parse(matrixInput); } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } async function checkDiskUsage(logger) { try { @@ -122228,9 +122338,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -122714,19 +122824,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -122742,9 +122852,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -122974,8 +123084,8 @@ var path4 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); @@ -123004,13 +123114,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -123213,7 +123323,7 @@ function formatDuration(durationMs) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { @@ -123347,6 +123457,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -123458,21 +123573,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, @@ -123768,6 +123883,8 @@ ${jsonContents}` var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, @@ -125646,9 +125763,9 @@ var JobStatus = /* @__PURE__ */ ((JobStatus2) => { JobStatus2["ConfigErrorStatus"] = "JOB_STATUS_CONFIGURATION_ERROR"; return JobStatus2; })(JobStatus || {}); -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -127281,15 +127398,15 @@ function validateSarifFileSchema(sarif, sarifFilePath, logger) { const warnings = (result.errors ?? []).filter( (err) => err.name === "format" && typeof err.argument === "string" && warningAttributes.includes(err.argument) ); - for (const warning11 of warnings) { + for (const warning12 of warnings) { logger.info( - `Warning: '${warning11.instance}' is not a valid URI in '${warning11.property}'.` + `Warning: '${warning12.instance}' is not a valid URI in '${warning12.property}'.` ); } if (errors.length > 0) { - for (const error4 of errors) { - logger.startGroup(`Error details: ${error4.stack}`); - logger.info(JSON.stringify(error4, null, 2)); + for (const error3 of errors) { + logger.startGroup(`Error details: ${error3.stack}`); + logger.info(JSON.stringify(error3, null, 2)); logger.endGroup(); } const sarifErrors = errors.map((e) => `- ${e.stack}`); @@ -127514,10 +127631,10 @@ function shouldConsiderConfigurationError(processingErrors) { } function shouldConsiderInvalidRequest(processingErrors) { return processingErrors.every( - (error4) => error4.startsWith("rejecting SARIF") || error4.startsWith("an invalid URI was provided as a SARIF location") || error4.startsWith("locationFromSarifResult: expected artifact location") || error4.startsWith( + (error3) => error3.startsWith("rejecting SARIF") || error3.startsWith("an invalid URI was provided as a SARIF location") || error3.startsWith("locationFromSarifResult: expected artifact location") || error3.startsWith( "could not convert rules: invalid security severity value, is not a number" ) || /^SARIF URI scheme [^\s]* did not match the checkout URI scheme [^\s]*/.test( - error4 + error3 ) ); } @@ -127728,8 +127845,8 @@ function getCheckoutPathInputOrThrow(workflow, jobName, matrixVars) { } // src/init-action-post-helper.ts -function createFailedUploadFailedSarifResult(error4) { - const wrappedError = wrapError(error4); +function createFailedUploadFailedSarifResult(error3) { + const wrappedError = wrapError(error3); return { upload_failed_run_error: wrappedError.message, upload_failed_run_stack_trace: wrappedError.stack @@ -127822,9 +127939,9 @@ async function run(uploadAllAvailableDebugArtifacts, printDebugLogs2, codeql, co ); } if (process.env["CODEQL_ACTION_EXPECT_UPLOAD_FAILED_SARIF"] === "true" && !uploadFailedSarifResult.raw_upload_size_bytes) { - const error4 = JSON.stringify(uploadFailedSarifResult); + const error3 = JSON.stringify(uploadFailedSarifResult); throw new Error( - `Expected to upload a failed SARIF file for this CodeQL code scanning run, but the result was instead ${error4}.` + `Expected to upload a failed SARIF file for this CodeQL code scanning run, but the result was instead ${error3}.` ); } if (process.env["CODEQL_ACTION_EXPECT_UPLOAD_FAILED_SARIF"] === "true") { @@ -127977,17 +128094,17 @@ async function runWrapper() { } } } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); - core17.setFailed(error4.message); + const error3 = wrapError(unwrappedError); + core17.setFailed(error3.message); const statusReportBase2 = await createStatusReportBase( "init-post" /* InitPost */, - getActionsStatus(error4), + getActionsStatus(error3), startedAt, config, await checkDiskUsage(logger), logger, - error4.message, - error4.stack + error3.message, + error3.stack ); if (statusReportBase2 !== void 0) { await sendStatusReport(statusReportBase2); @@ -128101,7 +128218,7 @@ tmp/lib/tmp.js: *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) long/index.js: (** diff --git a/lib/init-action.js b/lib/init-action.js index 8d63c95de..b5bf66216 100644 --- a/lib/init-action.js +++ b/lib/init-action.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug5("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve9(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug3() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug5; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning10(message, properties = {}) { + exports2.error = error3; + function warning11(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning10; + exports2.warning = warning11; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -23133,8 +23133,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -23866,7 +23866,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -23875,7 +23875,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -23885,17 +23885,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24514,7 +24514,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -24523,7 +24523,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -24533,17 +24533,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -27215,9 +27215,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { @@ -27662,23 +27662,21 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -27686,10 +27684,10 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", @@ -27699,9 +27697,9 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27723,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -27911,8 +27910,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27921,8 @@ var require_light = __commonJS({ return (await Promise.all(promises3)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28034,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28071,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run2, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28089,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } } doExpire(clearGlobalState, run2, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run2, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run2, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28116,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28395,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve9, returned, task; + var args, cb, error3, reject, resolve9, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve9, reject } = this._queue.shift()); @@ -28407,9 +28406,9 @@ var require_light = __commonJS({ return resolve9(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28542,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28876,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29179,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29212,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29237,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -34086,7 +34085,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core14 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob2 = __importStar4(require_glob()); var io7 = __importStar4(require_io3()); var crypto2 = __importStar4(require("crypto")); @@ -34170,7 +34169,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core14.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -36236,8 +36235,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36471,9 +36470,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37804,14 +37803,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37821,7 +37820,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37829,8 +37828,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -38050,7 +38049,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39260,14 +39259,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39939,11 +39938,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39958,8 +39957,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40453,8 +40452,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40688,9 +40687,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41190,8 +41189,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41425,9 +41424,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42490,9 +42489,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42540,13 +42539,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42566,21 +42565,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42745,8 +42744,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43152,16 +43151,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43705,10 +43704,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45787,12 +45786,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -46057,16 +46056,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46191,7 +46190,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46357,7 +46356,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46598,9 +46597,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47308,7 +47307,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48371,26 +48370,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48431,12 +48430,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48444,13 +48443,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48459,7 +48458,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -63065,8 +63064,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -63100,10 +63099,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64687,8 +64686,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64703,9 +64702,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve9, reject) => { this.emitter.on("finish", resolve9); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65806,8 +65805,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65891,7 +65890,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -69043,7 +69042,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70410,9 +70409,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70526,12 +70525,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70565,13 +70564,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71387,8 +71386,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73234,8 +73233,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74806,12 +74805,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74831,12 +74830,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75300,8 +75299,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75312,8 +75311,8 @@ var require_test_transport = __commonJS({ stream2.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } } @@ -76376,8 +76375,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76473,8 +76472,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76505,18 +76504,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76784,8 +76783,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76986,22 +76985,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -77056,15 +77055,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -77072,8 +77071,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -77135,10 +77134,10 @@ var require_cache3 = __commonJS({ } core14.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else { @@ -77151,8 +77150,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77197,8 +77196,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core14.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core14.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError2(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core14.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77217,10 +77216,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77235,8 +77234,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -81059,7 +81058,7 @@ var require_debug2 = __commonJS({ if (!debug5) { try { debug5 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug5 !== "function") { debug5 = function() { @@ -81092,8 +81091,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -81167,9 +81166,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -81336,10 +81335,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -81538,12 +81537,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -81642,7 +81641,6 @@ var fsPromises = __toESM(require("fs/promises")); var os = __toESM(require("os")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js @@ -81660,14 +81658,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs15.lstat(itemPath, { bigint: true }) : await fs15.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs15.lstat(itemPath, { bigint: true }) : await fs15.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs15.readdir(itemPath) : await fs15.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs15.readdir(itemPath) : await fs15.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -81678,13 +81676,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -82564,6 +82562,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -82683,7 +82693,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -82723,16 +82733,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -84307,9 +84308,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -84697,11 +84698,11 @@ function parseMatrixInput(matrixInput) { } return JSON.parse(matrixInput); } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } function prettyPrintPack(pack) { return `${pack.name}${pack.version ? `@${pack.version}` : ""}${pack.path ? `:${pack.path}` : ""}`; @@ -84727,9 +84728,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -84741,7 +84742,7 @@ function checkActionVersion(version, githubVersion) { semver.coerce(githubVersion.version) ?? "0.0.0", ">=3.20" )) { - core3.error( + core3.warning( "CodeQL Action v3 will be deprecated in December 2026. Please update all occurrences of the CodeQL Action in your workflow files to v4. For more information, see https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/" ); core3.exportVariable("CODEQL_ACTION_DID_LOG_VERSION_DEPRECATION" /* LOG_VERSION_DEPRECATION */, "true"); @@ -85306,9 +85307,9 @@ function getInvalidConfigFileMessage(configFile, messages) { return `The configuration file "${configFile}" is invalid: ${messages.slice(0, 10).join(", ")}${andMore}`; } function getConfigFileRepoFormatInvalidMessage(configFile) { - let error4 = `The configuration file "${configFile}" is not a supported remote file reference.`; - error4 += " Expected format //@"; - return error4; + let error3 = `The configuration file "${configFile}" is not a supported remote file reference.`; + error3 += " Expected format //@"; + return error3; } function getConfigFileFormatInvalidMessage(configFile) { return `The configuration file "${configFile}" could not be read`; @@ -85319,15 +85320,15 @@ function getConfigFileDirectoryGivenMessage(configFile) { function getEmptyCombinesError() { return `A '+' was used to specify that you want to add extra arguments to the configuration, but no extra arguments were specified. Please either remove the '+' or specify some extra arguments.`; } -function getConfigFilePropertyError(configFile, property, error4) { +function getConfigFilePropertyError(configFile, property, error3) { if (configFile === void 0) { - return `The workflow property "${property}" is invalid: ${error4}`; + return `The workflow property "${property}" is invalid: ${error3}`; } else { - return `The configuration file "${configFile}" is invalid: property "${property}" ${error4}`; + return `The configuration file "${configFile}" is invalid: property "${property}" ${error3}`; } } -function getRepoPropertyError(propertyName, error4) { - return `The repository property "${propertyName}" is invalid: ${error4}`; +function getRepoPropertyError(propertyName, error3) { + return `The repository property "${propertyName}" is invalid: ${error3}`; } function getPacksStrInvalid(packStr, configFile) { return configFile ? getConfigFilePropertyError( @@ -85606,8 +85607,8 @@ function parseUserConfig(logger, pathInput, contents, validateConfig) { if (validateConfig) { const result = new jsonschema.Validator().validate(doc, schema2); if (result.errors.length > 0) { - for (const error4 of result.errors) { - logger.error(error4.stack); + for (const error3 of result.errors) { + logger.error(error3.stack); } throw new ConfigurationError( getInvalidConfigFileMessage( @@ -85618,13 +85619,13 @@ function parseUserConfig(logger, pathInput, contents, validateConfig) { } } return doc; - } catch (error4) { - if (error4 instanceof YAMLException) { + } catch (error3) { + if (error3 instanceof YAMLException) { throw new ConfigurationError( - getConfigFileParseErrorMessage(pathInput, error4.message) + getConfigFileParseErrorMessage(pathInput, error3.message) ); } - throw error4; + throw error3; } } @@ -85634,8 +85635,8 @@ var path5 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); @@ -85664,13 +85665,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -85851,7 +85852,7 @@ function formatDuration(durationMs) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { @@ -86002,9 +86003,9 @@ async function downloadOverlayBaseDatabaseFromCache(codeql, config, logger) { logger.info( `Downloaded overlay-base database in cache with key ${foundKey}` ); - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to download overlay-base database from cache: ${error4 instanceof Error ? error4.message : String(error4)}` + `Failed to download overlay-base database from cache: ${error3 instanceof Error ? error3.message : String(error3)}` ); return void 0; } @@ -86110,6 +86111,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -86221,21 +86227,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, @@ -86655,6 +86661,8 @@ async function cachePrefix(codeql, language) { } // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; async function getSupportedLanguageMap(codeql, logger) { const resolveSupportedLanguagesUsingCli = await codeql.supportsFeature( "builtinExtractorsSpecifyDefaultQueries" /* BuiltinExtractorsSpecifyDefaultQueries */ @@ -86829,6 +86837,7 @@ async function initActionState({ trapCaches, trapCacheDownloadTime, dependencyCachingEnabled: getCachingKind(dependencyCachingEnabled), + dependencyCachingRestoredKeys: [], extraQueryExclusions: [], overlayDatabaseMode: "none" /* None */, useOverlayDatabaseCaching: false, @@ -86889,10 +86898,7 @@ var OVERLAY_ANALYSIS_CODE_SCANNING_FEATURES = { rust: "overlay_analysis_code_scanning_rust" /* OverlayAnalysisCodeScanningRust */, swift: "overlay_analysis_code_scanning_swift" /* OverlayAnalysisCodeScanningSwift */ }; -async function isOverlayAnalysisFeatureEnabled(repository, features, codeql, languages, codeScanningConfig) { - if (!["github", "dsp-testing"].includes(repository.owner)) { - return false; - } +async function isOverlayAnalysisFeatureEnabled(features, codeql, languages, codeScanningConfig) { if (!await features.getValue("overlay_analysis" /* OverlayAnalysis */, codeql)) { return false; } @@ -86914,7 +86920,7 @@ async function isOverlayAnalysisFeatureEnabled(repository, features, codeql, lan } return true; } -async function getOverlayDatabaseMode(codeql, repository, features, languages, sourceRoot, buildMode, codeScanningConfig, logger) { +async function getOverlayDatabaseMode(codeql, features, languages, sourceRoot, buildMode, codeScanningConfig, logger) { let overlayDatabaseMode = "none" /* None */; let useOverlayDatabaseCaching = false; const modeEnv = process.env.CODEQL_OVERLAY_DATABASE_MODE; @@ -86924,24 +86930,33 @@ async function getOverlayDatabaseMode(codeql, repository, features, languages, s `Setting overlay database mode to ${overlayDatabaseMode} from the CODEQL_OVERLAY_DATABASE_MODE environment variable.` ); } else if (await isOverlayAnalysisFeatureEnabled( - repository, features, codeql, languages, codeScanningConfig )) { - if (isAnalyzingPullRequest()) { - overlayDatabaseMode = "overlay" /* Overlay */; - useOverlayDatabaseCaching = true; + const diskUsage = await checkDiskUsage(logger); + if (diskUsage === void 0 || diskUsage.numAvailableBytes < OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES) { + const diskSpaceMb = diskUsage === void 0 ? 0 : Math.round(diskUsage.numAvailableBytes / 1e6); + overlayDatabaseMode = "none" /* None */; + useOverlayDatabaseCaching = false; logger.info( - `Setting overlay database mode to ${overlayDatabaseMode} with caching because we are analyzing a pull request.` - ); - } else if (await isAnalyzingDefaultBranch()) { - overlayDatabaseMode = "overlay-base" /* OverlayBase */; - useOverlayDatabaseCaching = true; - logger.info( - `Setting overlay database mode to ${overlayDatabaseMode} with caching because we are analyzing the default branch.` + `Setting overlay database mode to ${overlayDatabaseMode} due to insufficient disk space (${diskSpaceMb} MB).` ); + } else { + if (isAnalyzingPullRequest()) { + overlayDatabaseMode = "overlay" /* Overlay */; + useOverlayDatabaseCaching = true; + logger.info( + `Setting overlay database mode to ${overlayDatabaseMode} with caching because we are analyzing a pull request.` + ); + } else if (await isAnalyzingDefaultBranch()) { + overlayDatabaseMode = "overlay-base" /* OverlayBase */; + useOverlayDatabaseCaching = true; + logger.info( + `Setting overlay database mode to ${overlayDatabaseMode} with caching because we are analyzing the default branch.` + ); + } } } const nonOverlayAnalysis = { @@ -87032,7 +87047,6 @@ async function initConfig(features, inputs) { } const { overlayDatabaseMode, useOverlayDatabaseCaching } = await getOverlayDatabaseMode( inputs.codeql, - inputs.repository, inputs.features, config.languages, inputs.sourceRoot, @@ -87340,6 +87354,7 @@ async function checkHashPatterns(codeql, features, language, cacheConfig, checkT } async function downloadDependencyCaches(codeql, features, languages, logger) { const status = []; + const restoredKeys = []; for (const language of languages) { const cacheConfig = defaultCacheConfigs[language]; if (cacheConfig === void 0) { @@ -87378,14 +87393,22 @@ async function downloadDependencyCaches(codeql, features, languages, logger) { const download_duration_ms = Math.round(performance.now() - start); if (hitKey !== void 0) { logger.info(`Cache hit on key ${hitKey} for ${language}.`); - const hit_kind = hitKey === primaryKey ? "exact" /* Exact */ : "partial" /* Partial */; - status.push({ language, hit_kind, download_duration_ms }); + let hit_kind = "partial" /* Partial */; + if (hitKey === primaryKey) { + hit_kind = "exact" /* Exact */; + } + status.push({ + language, + hit_kind, + download_duration_ms + }); + restoredKeys.push(hitKey); } else { status.push({ language, hit_kind: "miss" /* Miss */ }); logger.info(`No suitable cache found for ${language}.`); } } - return status; + return { statusReport: status, restoredKeys }; } async function cacheKey2(codeql, features, language, patterns) { const hash = await glob.hashFiles(patterns.join("\n")); @@ -87542,19 +87565,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -87570,9 +87593,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -89403,9 +89426,9 @@ function isFirstPartyAnalysis(actionName) { } return process.env["CODEQL_ACTION_INIT_HAS_RUN" /* INIT_ACTION_HAS_RUN */] === "true"; } -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -89827,16 +89850,16 @@ async function sendStartingStatusReport(startedAt, config, logger) { await sendStatusReport(statusReportBase); } } -async function sendCompletedStatusReport(startedAt, config, configFile, toolsDownloadStatusReport, toolsFeatureFlagsValid, toolsSource, toolsVersion, overlayBaseDatabaseStats, dependencyCachingResults, logger, error4) { +async function sendCompletedStatusReport(startedAt, config, configFile, toolsDownloadStatusReport, toolsFeatureFlagsValid, toolsSource, toolsVersion, overlayBaseDatabaseStats, dependencyCachingResults, logger, error3) { const statusReportBase = await createStatusReportBase( "init" /* Init */, - getActionsStatus(error4), + getActionsStatus(error3), startedAt, config, await checkDiskUsage(logger), logger, - error4?.message, - error4?.stack + error3?.message, + error3?.stack ); if (statusReportBase === void 0) { return; @@ -90000,17 +90023,17 @@ async function run() { }); await checkInstallPython311(config.languages, codeql); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); - core13.setFailed(error4.message); + const error3 = wrapError(unwrappedError); + core13.setFailed(error3.message); const statusReportBase = await createStatusReportBase( "init" /* Init */, - error4 instanceof ConfigurationError ? "user-error" : "aborted", + error3 instanceof ConfigurationError ? "user-error" : "aborted", startedAt, config, await checkDiskUsage(logger), logger, - error4.message, - error4.stack + error3.message, + error3.stack ); if (statusReportBase !== void 0) { await sendStatusReport(statusReportBase); @@ -90018,7 +90041,7 @@ async function run() { return; } let overlayBaseDatabaseStats; - let dependencyCachingResults; + let dependencyCachingStatus; try { if (config.overlayDatabaseMode === "overlay" /* Overlay */ && config.useOverlayDatabaseCaching) { overlayBaseDatabaseStats = await downloadOverlayBaseDatabaseFromCache( @@ -90159,12 +90182,14 @@ exec ${goBinaryPath} "$@"` } } if (shouldRestoreCache(config.dependencyCachingEnabled)) { - dependencyCachingResults = await downloadDependencyCaches( + const dependencyCachingResult = await downloadDependencyCaches( codeql, features, config.languages, logger ); + dependencyCachingStatus = dependencyCachingResult.statusReport; + config.dependencyCachingRestoredKeys = dependencyCachingResult.restoredKeys; } if (await codeQlVersionAtLeast(codeql, "2.17.1")) { } else { @@ -90253,8 +90278,8 @@ exec ${goBinaryPath} "$@"` core13.setOutput("codeql-path", config.codeQLCmd); core13.setOutput("codeql-version", (await codeql.getVersion()).version); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); - core13.setFailed(error4.message); + const error3 = wrapError(unwrappedError); + core13.setFailed(error3.message); await sendCompletedStatusReport( startedAt, config, @@ -90265,9 +90290,9 @@ exec ${goBinaryPath} "$@"` toolsSource, toolsVersion, overlayBaseDatabaseStats, - dependencyCachingResults, + dependencyCachingStatus, logger, - error4 + error3 ); return; } finally { @@ -90283,7 +90308,7 @@ exec ${goBinaryPath} "$@"` toolsSource, toolsVersion, overlayBaseDatabaseStats, - dependencyCachingResults, + dependencyCachingStatus, logger ); } @@ -90316,8 +90341,8 @@ async function recordZstdAvailability(config, zstdAvailability) { async function runWrapper() { try { await run(); - } catch (error4) { - core13.setFailed(`init action failed: ${getErrorMessage(error4)}`); + } catch (error3) { + core13.setFailed(`init action failed: ${getErrorMessage(error3)}`); } await checkForTimeout(); } @@ -90331,5 +90356,5 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/resolve-environment-action.js b/lib/resolve-environment-action.js index 766a59c17..96e60bb69 100644 --- a/lib/resolve-environment-action.js +++ b/lib/resolve-environment-action.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug5("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve4(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug5; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning8(message, properties = {}) { + exports2.error = error3; + function warning9(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning8; + exports2.warning = warning9; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { @@ -27662,23 +27662,21 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -27686,10 +27684,10 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", @@ -27699,9 +27697,9 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27723,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -27911,8 +27910,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27921,8 @@ var require_light = __commonJS({ return (await Promise.all(promises2)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28034,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28071,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run2, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28089,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } } doExpire(clearGlobalState, run2, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run2, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run2, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28116,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28395,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve4, returned, task; + var args, cb, error3, reject, resolve4, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve4, reject } = this._queue.shift()); @@ -28407,9 +28406,9 @@ var require_light = __commonJS({ return resolve4(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28542,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28876,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29179,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29212,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29237,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -33935,7 +33934,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core13 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob = __importStar4(require_glob()); var io5 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -34019,7 +34018,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core13.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -36085,8 +36084,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36320,9 +36319,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37653,14 +37652,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37670,7 +37669,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37678,8 +37677,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -37899,7 +37898,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39109,14 +39108,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39788,11 +39787,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39807,8 +39806,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40302,8 +40301,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40537,9 +40536,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41039,8 +41038,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41274,9 +41273,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42339,9 +42338,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42389,13 +42388,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42415,21 +42414,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42594,8 +42593,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43001,16 +43000,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43554,10 +43553,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45636,12 +45635,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -45906,16 +45905,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46040,7 +46039,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46206,7 +46205,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46447,9 +46446,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47157,7 +47156,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48220,26 +48219,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48280,12 +48279,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48293,13 +48292,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48308,7 +48307,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -62914,8 +62913,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -62949,10 +62948,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64536,8 +64535,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64552,9 +64551,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve4, reject) => { this.emitter.on("finish", resolve4); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65655,8 +65654,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65740,7 +65739,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -68892,7 +68891,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70259,9 +70258,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core13.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core13.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70375,12 +70374,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70414,13 +70413,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71236,8 +71235,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73083,8 +73082,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74655,12 +74654,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74680,12 +74679,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75149,8 +75148,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75161,8 +75160,8 @@ var require_test_transport = __commonJS({ stream.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } } @@ -76225,8 +76224,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76322,8 +76321,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76354,18 +76353,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76633,8 +76632,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76835,22 +76834,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core13.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core13.error(`Failed to restore: ${error4.message}`); + core13.error(`Failed to restore: ${error3.message}`); } else { - core13.warning(`Failed to restore: ${error4.message}`); + core13.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76905,15 +76904,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core13.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core13.error(`Failed to restore: ${error4.message}`); + core13.error(`Failed to restore: ${error3.message}`); } else { - core13.warning(`Failed to restore: ${error4.message}`); + core13.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -76921,8 +76920,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76984,10 +76983,10 @@ var require_cache3 = __commonJS({ } core13.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core13.info(`Failed to save: ${typedError.message}`); } else { @@ -77000,8 +76999,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77046,8 +77045,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core13.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core13.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core13.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77066,10 +77065,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core13.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77084,8 +77083,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79810,7 @@ var require_debug2 = __commonJS({ if (!debug5) { try { debug5 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug5 !== "function") { debug5 = function() { @@ -79844,8 +79843,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79918,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80087,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80289,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -80333,7 +80332,6 @@ var io2 = __toESM(require_io2()); var fsPromises = __toESM(require("fs/promises")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js @@ -80351,14 +80349,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs5.lstat(itemPath, { bigint: true }) : await fs5.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs5.lstat(itemPath, { bigint: true }) : await fs5.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs5.readdir(itemPath) : await fs5.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs5.readdir(itemPath) : await fs5.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -80369,13 +80367,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -81255,6 +81253,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -81374,7 +81384,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -81414,16 +81424,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -82993,9 +82994,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -83135,11 +83136,11 @@ async function checkForTimeout() { process.exit(); } } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } async function checkDiskUsage(logger) { try { @@ -83162,9 +83163,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -83176,7 +83177,7 @@ function checkActionVersion(version, githubVersion) { semver.coerce(githubVersion.version) ?? "0.0.0", ">=3.20" )) { - core3.error( + core3.warning( "CodeQL Action v3 will be deprecated in December 2026. Please update all occurrences of the CodeQL Action in your workflow files to v4. For more information, see https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/" ); core3.exportVariable("CODEQL_ACTION_DID_LOG_VERSION_DEPRECATION" /* LOG_VERSION_DEPRECATION */, "true"); @@ -83458,19 +83459,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -83486,9 +83487,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -83723,13 +83724,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -83879,7 +83880,7 @@ function getActionsLogger() { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { @@ -84006,6 +84007,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -84117,21 +84123,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, @@ -84144,6 +84150,8 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, @@ -84814,9 +84822,9 @@ function isFirstPartyAnalysis(actionName) { } return process.env["CODEQL_ACTION_INIT_HAS_RUN" /* INIT_ACTION_HAS_RUN */] === "true"; } -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -85022,25 +85030,25 @@ async function run() { ); core12.setOutput(ENVIRONMENT_OUTPUT_NAME, result); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); - if (error4 instanceof CliError) { + const error3 = wrapError(unwrappedError); + if (error3 instanceof CliError) { core12.setOutput(ENVIRONMENT_OUTPUT_NAME, {}); logger.warning( - `Failed to resolve a build environment suitable for automatically building your code. ${error4.message}` + `Failed to resolve a build environment suitable for automatically building your code. ${error3.message}` ); } else { core12.setFailed( - `Failed to resolve a build environment suitable for automatically building your code. ${error4.message}` + `Failed to resolve a build environment suitable for automatically building your code. ${error3.message}` ); const statusReportBase2 = await createStatusReportBase( "resolve-environment" /* ResolveEnvironment */, - getActionsStatus(error4), + getActionsStatus(error3), startedAt, config, await checkDiskUsage(logger), logger, - error4.message, - error4.stack + error3.message, + error3.stack ); if (statusReportBase2 !== void 0) { await sendStatusReport(statusReportBase2); @@ -85063,10 +85071,10 @@ async function run() { async function runWrapper() { try { await run(); - } catch (error4) { + } catch (error3) { core12.setFailed( `${"resolve-environment" /* ResolveEnvironment */} action failed: ${getErrorMessage( - error4 + error3 )}` ); } @@ -85082,5 +85090,5 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/setup-codeql-action.js b/lib/setup-codeql-action.js index f00d601e7..06df5ebf8 100644 --- a/lib/setup-codeql-action.js +++ b/lib/setup-codeql-action.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug5("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve4(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug5; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning8(message, properties = {}) { + exports2.error = error3; + function warning9(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning8; + exports2.warning = warning9; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { @@ -27662,23 +27662,21 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -27686,10 +27684,10 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", @@ -27699,9 +27697,9 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27723,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -27911,8 +27910,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27921,8 @@ var require_light = __commonJS({ return (await Promise.all(promises3)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28034,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28071,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run2, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28089,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } } doExpire(clearGlobalState, run2, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run2, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run2, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28116,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28395,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve4, returned, task; + var args, cb, error3, reject, resolve4, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve4, reject } = this._queue.shift()); @@ -28407,9 +28406,9 @@ var require_light = __commonJS({ return resolve4(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28542,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28876,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29179,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29212,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29237,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -32638,7 +32637,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core13 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob = __importStar4(require_glob()); var io6 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -32722,7 +32721,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core13.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -34788,8 +34787,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -35023,9 +35022,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -36356,14 +36355,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -36373,7 +36372,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -36381,8 +36380,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -36602,7 +36601,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -37812,14 +37811,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -38491,11 +38490,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -38510,8 +38509,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -39005,8 +39004,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -39240,9 +39239,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -39742,8 +39741,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -39977,9 +39976,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -41042,9 +41041,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -41092,13 +41091,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -41118,21 +41117,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -41297,8 +41296,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -41704,16 +41703,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -42257,10 +42256,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -44339,12 +44338,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -44609,16 +44608,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -44743,7 +44742,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -44909,7 +44908,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -45150,9 +45149,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -45860,7 +45859,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -46923,26 +46922,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -46983,12 +46982,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -46996,13 +46995,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -47011,7 +47010,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -61617,8 +61616,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -61652,10 +61651,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -63239,8 +63238,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -63255,9 +63254,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve4, reject) => { this.emitter.on("finish", resolve4); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -64358,8 +64357,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -64443,7 +64442,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -67595,7 +67594,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -68962,9 +68961,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core13.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core13.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -69078,12 +69077,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -69117,13 +69116,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -69939,8 +69938,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -71786,8 +71785,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -73358,12 +73357,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -73383,12 +73382,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -73852,8 +73851,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -73864,8 +73863,8 @@ var require_test_transport = __commonJS({ stream2.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } } @@ -74928,8 +74927,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -75025,8 +75024,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -75057,18 +75056,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -75336,8 +75335,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -75538,22 +75537,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core13.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core13.error(`Failed to restore: ${error4.message}`); + core13.error(`Failed to restore: ${error3.message}`); } else { - core13.warning(`Failed to restore: ${error4.message}`); + core13.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -75608,15 +75607,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core13.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core13.error(`Failed to restore: ${error4.message}`); + core13.error(`Failed to restore: ${error3.message}`); } else { - core13.warning(`Failed to restore: ${error4.message}`); + core13.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -75624,8 +75623,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -75687,10 +75686,10 @@ var require_cache3 = __commonJS({ } core13.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core13.info(`Failed to save: ${typedError.message}`); } else { @@ -75703,8 +75702,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -75749,8 +75748,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core13.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core13.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core13.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -75769,10 +75768,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core13.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -75787,8 +75786,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core13.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core13.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79810,7 @@ var require_debug2 = __commonJS({ if (!debug5) { try { debug5 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug5 !== "function") { debug5 = function() { @@ -79844,8 +79843,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79918,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80087,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80289,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -80389,7 +80388,6 @@ var fs = __toESM(require("fs")); var fsPromises = __toESM(require("fs/promises")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js @@ -80407,14 +80405,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs9.lstat(itemPath, { bigint: true }) : await fs9.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs9.lstat(itemPath, { bigint: true }) : await fs9.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs9.readdir(itemPath) : await fs9.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs9.readdir(itemPath) : await fs9.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -80425,13 +80423,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -81311,6 +81309,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -81430,7 +81440,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -81470,16 +81480,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -83050,9 +83051,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -83211,11 +83212,11 @@ async function checkForTimeout() { process.exit(); } } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } async function checkDiskUsage(logger) { try { @@ -83238,9 +83239,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -83252,7 +83253,7 @@ function checkActionVersion(version, githubVersion) { semver.coerce(githubVersion.version) ?? "0.0.0", ">=3.20" )) { - core3.error( + core3.warning( "CodeQL Action v3 will be deprecated in December 2026. Please update all occurrences of the CodeQL Action in your workflow files to v4. For more information, see https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/" ); core3.exportVariable("CODEQL_ACTION_DID_LOG_VERSION_DEPRECATION" /* LOG_VERSION_DEPRECATION */, "true"); @@ -83588,8 +83589,8 @@ var path4 = __toESM(require("path")); var semver3 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); @@ -83621,13 +83622,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -83788,7 +83789,7 @@ function formatDuration(durationMs) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { @@ -83918,6 +83919,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -84029,21 +84035,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, @@ -84357,19 +84363,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -84385,9 +84391,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -84589,6 +84595,8 @@ var PACK_IDENTIFIER_PATTERN = (function() { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, @@ -86135,9 +86143,9 @@ function isFirstPartyAnalysis(actionName) { } return process.env["CODEQL_ACTION_INIT_HAS_RUN" /* INIT_ACTION_HAS_RUN */] === "true"; } -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -86308,16 +86316,16 @@ async function sendStatusReport(statusReport) { } // src/setup-codeql-action.ts -async function sendCompletedStatusReport(startedAt, toolsDownloadStatusReport, toolsFeatureFlagsValid, toolsSource, toolsVersion, logger, error4) { +async function sendCompletedStatusReport(startedAt, toolsDownloadStatusReport, toolsFeatureFlagsValid, toolsSource, toolsVersion, logger, error3) { const statusReportBase = await createStatusReportBase( "setup-codeql" /* SetupCodeQL */, - getActionsStatus(error4), + getActionsStatus(error3), startedAt, void 0, await checkDiskUsage(logger), logger, - error4?.message, - error4?.stack + error3?.message, + error3?.stack ); if (statusReportBase === void 0) { return; @@ -86399,17 +86407,17 @@ async function run() { core12.setOutput("codeql-version", (await codeql.getVersion()).version); core12.exportVariable("CODEQL_ACTION_SETUP_CODEQL_HAS_RUN" /* SETUP_CODEQL_ACTION_HAS_RUN */, "true"); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); - core12.setFailed(error4.message); + const error3 = wrapError(unwrappedError); + core12.setFailed(error3.message); const statusReportBase = await createStatusReportBase( "setup-codeql" /* SetupCodeQL */, - error4 instanceof ConfigurationError ? "user-error" : "failure", + error3 instanceof ConfigurationError ? "user-error" : "failure", startedAt, void 0, await checkDiskUsage(logger), logger, - error4.message, - error4.stack + error3.message, + error3.stack ); if (statusReportBase !== void 0) { await sendStatusReport(statusReportBase); @@ -86428,8 +86436,8 @@ async function run() { async function runWrapper() { try { await run(); - } catch (error4) { - core12.setFailed(`setup-codeql action failed: ${getErrorMessage(error4)}`); + } catch (error3) { + core12.setFailed(`setup-codeql action failed: ${getErrorMessage(error3)}`); } await checkForTimeout(); } @@ -86443,5 +86451,5 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/start-proxy-action-post.js b/lib/start-proxy-action-post.js index 85e0aaeb0..c3fcc799e 100644 --- a/lib/start-proxy-action-post.js +++ b/lib/start-proxy-action-post.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug4("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay === "number" && delay > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve2(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed; function isDebug3() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug4; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning8(message, properties = {}) { + exports2.error = error3; + function warning9(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning8; + exports2.warning = warning9; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { @@ -27662,23 +27662,21 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -27686,10 +27684,10 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", @@ -27699,9 +27697,9 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27723,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -27911,8 +27910,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27921,8 @@ var require_light = __commonJS({ return (await Promise.all(promises)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28034,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28071,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28089,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } } doExpire(clearGlobalState, run, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28116,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28395,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve2, returned, task; + var args, cb, error3, reject, resolve2, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve2, reject } = this._queue.shift()); @@ -28407,9 +28406,9 @@ var require_light = __commonJS({ return resolve2(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28542,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28876,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29179,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29212,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29237,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -33935,7 +33934,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core14 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob2 = __importStar4(require_glob()); var io6 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -34019,7 +34018,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core14.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -36085,8 +36084,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36320,9 +36319,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37653,14 +37652,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37670,7 +37669,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37678,8 +37677,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -37899,7 +37898,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39109,14 +39108,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39788,11 +39787,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39807,8 +39806,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40302,8 +40301,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40537,9 +40536,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41039,8 +41038,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41274,9 +41273,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42339,9 +42338,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42389,13 +42388,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42415,21 +42414,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42594,8 +42593,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43001,16 +43000,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43554,10 +43553,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45636,12 +45635,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -45906,16 +45905,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46040,7 +46039,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46206,7 +46205,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46447,9 +46446,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47157,7 +47156,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48220,26 +48219,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48280,12 +48279,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48293,13 +48292,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48308,7 +48307,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -62914,8 +62913,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -62949,10 +62948,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64536,8 +64535,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64552,9 +64551,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve2, reject) => { this.emitter.on("finish", resolve2); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65655,8 +65654,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65740,7 +65739,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -68892,7 +68891,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70259,9 +70258,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70375,12 +70374,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70414,13 +70413,13 @@ var require_requestUtils = __commonJS({ delay, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71236,8 +71235,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73083,8 +73082,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74655,12 +74654,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74680,12 +74679,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75149,8 +75148,8 @@ var require_test_transport = __commonJS({ } try { yield delay(this.responseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75161,8 +75160,8 @@ var require_test_transport = __commonJS({ stream.notifyMessage(msg); try { yield delay(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } } @@ -76225,8 +76224,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76322,8 +76321,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76354,18 +76353,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76633,8 +76632,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76835,22 +76834,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76905,15 +76904,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -76921,8 +76920,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76984,10 +76983,10 @@ var require_cache3 = __commonJS({ } core14.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else { @@ -77000,8 +76999,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77046,8 +77045,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core14.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core14.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError2(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core14.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77066,10 +77065,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77084,8 +77083,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -80413,8 +80412,8 @@ var require_util11 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } function maskSecretUrls(body) { @@ -80507,8 +80506,8 @@ var require_artifact_twirp_client2 = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -80539,18 +80538,18 @@ var require_artifact_twirp_client2 = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -80821,11 +80820,11 @@ var require_blob_upload = __commonJS({ blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options), chunkTimer((0, config_1.getUploadChunkTimeout)()) ]); - } catch (error4) { - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + } catch (error3) { + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } - throw error4; + throw error3; } finally { abortController.abort(); } @@ -81846,9 +81845,9 @@ var require_async = __commonJS({ invokeCallback(callback, err && (err instanceof Error || err.message) ? err : new Error(err)); }); } - function invokeCallback(callback, error4, value) { + function invokeCallback(callback, error3, value) { try { - callback(error4, value); + callback(error3, value); } catch (err) { setImmediate$1((e) => { throw e; @@ -83154,10 +83153,10 @@ var require_async = __commonJS({ function reflect(fn) { var _fn = wrapAsync(fn); return initialParams(function reflectOn(args, reflectCallback) { - args.push((error4, ...cbArgs) => { + args.push((error3, ...cbArgs) => { let retVal = {}; - if (error4) { - retVal.error = error4; + if (error3) { + retVal.error = error3; } if (cbArgs.length > 0) { var value = cbArgs; @@ -83313,13 +83312,13 @@ var require_async = __commonJS({ var timer; function timeoutCallback() { var name = asyncFn.name || "anonymous"; - var error4 = new Error('Callback function "' + name + '" timed out.'); - error4.code = "ETIMEDOUT"; + var error3 = new Error('Callback function "' + name + '" timed out.'); + error3.code = "ETIMEDOUT"; if (info7) { - error4.info = info7; + error3.info = info7; } timedOut = true; - callback(error4); + callback(error3); } args.push((...cbArgs) => { if (!timedOut) { @@ -83362,7 +83361,7 @@ var require_async = __commonJS({ return callback[PROMISE_SYMBOL]; } function tryEach(tasks, callback) { - var error4 = null; + var error3 = null; var result; return eachSeries$1(tasks, (task, taskCb) => { wrapAsync(task)((err, ...args) => { @@ -83372,10 +83371,10 @@ var require_async = __commonJS({ } else { result = args; } - error4 = err; + error3 = err; taskCb(err ? null : {}); }); - }, () => callback(error4, result)); + }, () => callback(error3, result)); } var tryEach$1 = awaitify(tryEach); function unmemoize(fn) { @@ -89855,19 +89854,19 @@ var require_from = __commonJS({ next(); } }; - readable._destroy = function(error4, cb) { + readable._destroy = function(error3, cb) { PromisePrototypeThen( - close(error4), - () => process2.nextTick(cb, error4), + close(error3), + () => process2.nextTick(cb, error3), // nextTick is here in case cb throws - (e) => process2.nextTick(cb, e || error4) + (e) => process2.nextTick(cb, e || error3) ); }; - async function close(error4) { - const hadError = error4 !== void 0 && error4 !== null; + async function close(error3) { + const hadError = error3 !== void 0 && error3 !== null; const hasThrow = typeof iterator.throw === "function"; if (hadError && hasThrow) { - const { value, done } = await iterator.throw(error4); + const { value, done } = await iterator.throw(error3); await value; if (done) { return; @@ -90075,12 +90074,12 @@ var require_readable3 = __commonJS({ this.destroy(err); }; Readable.prototype[SymbolAsyncDispose] = function() { - let error4; + let error3; if (!this.destroyed) { - error4 = this.readableEnded ? null : new AbortError(); - this.destroy(error4); + error3 = this.readableEnded ? null : new AbortError(); + this.destroy(error3); } - return new Promise2((resolve2, reject) => eos(this, (err) => err && err !== error4 ? reject(err) : resolve2(null))); + return new Promise2((resolve2, reject) => eos(this, (err) => err && err !== error3 ? reject(err) : resolve2(null))); }; Readable.prototype.push = function(chunk, encoding) { return readableAddChunk(this, chunk, encoding, false); @@ -90633,14 +90632,14 @@ var require_readable3 = __commonJS({ } } stream.on("readable", next); - let error4; + let error3; const cleanup = eos( stream, { writable: false }, (err) => { - error4 = err ? aggregateTwoErrors(error4, err) : null; + error3 = err ? aggregateTwoErrors(error3, err) : null; callback(); callback = nop; } @@ -90650,19 +90649,19 @@ var require_readable3 = __commonJS({ const chunk = stream.destroyed ? null : stream.read(); if (chunk !== null) { yield chunk; - } else if (error4) { - throw error4; - } else if (error4 === null) { + } else if (error3) { + throw error3; + } else if (error3 === null) { return; } else { await new Promise2(next); } } } catch (err) { - error4 = aggregateTwoErrors(error4, err); - throw error4; + error3 = aggregateTwoErrors(error3, err); + throw error3; } finally { - if ((error4 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error4 === void 0 || stream._readableState.autoDestroy)) { + if ((error3 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error3 === void 0 || stream._readableState.autoDestroy)) { destroyImpl.destroyer(stream, null); } else { stream.off("readable", next); @@ -92155,11 +92154,11 @@ var require_pipeline3 = __commonJS({ yield* Readable.prototype[SymbolAsyncIterator].call(val2); } async function pumpToNode(iterable, writable, finish, { end }) { - let error4; + let error3; let onresolve = null; const resume = (err) => { if (err) { - error4 = err; + error3 = err; } if (onresolve) { const callback = onresolve; @@ -92168,12 +92167,12 @@ var require_pipeline3 = __commonJS({ } }; const wait = () => new Promise2((resolve2, reject) => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { onresolve = () => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve2(); } @@ -92203,7 +92202,7 @@ var require_pipeline3 = __commonJS({ } finish(); } catch (err) { - finish(error4 !== err ? aggregateTwoErrors(error4, err) : err); + finish(error3 !== err ? aggregateTwoErrors(error3, err) : err); } finally { cleanup(); writable.off("drain", resume); @@ -92257,7 +92256,7 @@ var require_pipeline3 = __commonJS({ if (outerSignal) { disposable = addAbortListener(outerSignal, abort); } - let error4; + let error3; let value; const destroys = []; let finishCount = 0; @@ -92266,23 +92265,23 @@ var require_pipeline3 = __commonJS({ } function finishImpl(err, final) { var _disposable; - if (err && (!error4 || error4.code === "ERR_STREAM_PREMATURE_CLOSE")) { - error4 = err; + if (err && (!error3 || error3.code === "ERR_STREAM_PREMATURE_CLOSE")) { + error3 = err; } - if (!error4 && !final) { + if (!error3 && !final) { return; } while (destroys.length) { - destroys.shift()(error4); + destroys.shift()(error3); } ; (_disposable = disposable) === null || _disposable === void 0 ? void 0 : _disposable[SymbolDispose](); ac.abort(); if (final) { - if (!error4) { + if (!error3) { lastStreamCleanup.forEach((fn) => fn()); } - process2.nextTick(callback, error4, value); + process2.nextTick(callback, error3, value); } } let ret; @@ -94132,52 +94131,128 @@ var require_isPlainObject = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/brace-expansion/index.js -var require_brace_expansion3 = __commonJS({ - "node_modules/archiver-utils/node_modules/brace-expansion/index.js"(exports2, module2) { - var balanced = require_balanced_match(); - module2.exports = expandTop; +// node_modules/@isaacs/balanced-match/dist/commonjs/index.js +var require_commonjs13 = __commonJS({ + "node_modules/@isaacs/balanced-match/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.range = exports2.balanced = void 0; + var balanced = (a, b, str2) => { + const ma = a instanceof RegExp ? maybeMatch(a, str2) : a; + const mb = b instanceof RegExp ? maybeMatch(b, str2) : b; + const r = ma !== null && mb != null && (0, exports2.range)(ma, mb, str2); + return r && { + start: r[0], + end: r[1], + pre: str2.slice(0, r[0]), + body: str2.slice(r[0] + ma.length, r[1]), + post: str2.slice(r[1] + mb.length) + }; + }; + exports2.balanced = balanced; + var maybeMatch = (reg, str2) => { + const m = str2.match(reg); + return m ? m[0] : null; + }; + var range = (a, b, str2) => { + let begs, beg, left, right = void 0, result; + let ai = str2.indexOf(a); + let bi = str2.indexOf(b, ai + 1); + let i = ai; + if (ai >= 0 && bi > 0) { + if (a === b) { + return [ai, bi]; + } + begs = []; + left = str2.length; + while (i >= 0 && !result) { + if (i === ai) { + begs.push(i); + ai = str2.indexOf(a, i + 1); + } else if (begs.length === 1) { + const r = begs.pop(); + if (r !== void 0) + result = [r, bi]; + } else { + beg = begs.pop(); + if (beg !== void 0 && beg < left) { + left = beg; + right = bi; + } + bi = str2.indexOf(b, i + 1); + } + i = ai < bi && ai >= 0 ? ai : bi; + } + if (begs.length && right !== void 0) { + result = [left, right]; + } + } + return result; + }; + exports2.range = range; + } +}); + +// node_modules/@isaacs/brace-expansion/dist/commonjs/index.js +var require_commonjs14 = __commonJS({ + "node_modules/@isaacs/brace-expansion/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.expand = expand; + var balanced_match_1 = require_commonjs13(); var escSlash = "\0SLASH" + Math.random() + "\0"; var escOpen = "\0OPEN" + Math.random() + "\0"; var escClose = "\0CLOSE" + Math.random() + "\0"; var escComma = "\0COMMA" + Math.random() + "\0"; var escPeriod = "\0PERIOD" + Math.random() + "\0"; + var escSlashPattern = new RegExp(escSlash, "g"); + var escOpenPattern = new RegExp(escOpen, "g"); + var escClosePattern = new RegExp(escClose, "g"); + var escCommaPattern = new RegExp(escComma, "g"); + var escPeriodPattern = new RegExp(escPeriod, "g"); + var slashPattern = /\\\\/g; + var openPattern = /\\{/g; + var closePattern = /\\}/g; + var commaPattern = /\\,/g; + var periodPattern = /\\./g; function numeric(str2) { - return parseInt(str2, 10) == str2 ? parseInt(str2, 10) : str2.charCodeAt(0); + return !isNaN(str2) ? parseInt(str2, 10) : str2.charCodeAt(0); } function escapeBraces(str2) { - return str2.split("\\\\").join(escSlash).split("\\{").join(escOpen).split("\\}").join(escClose).split("\\,").join(escComma).split("\\.").join(escPeriod); + return str2.replace(slashPattern, escSlash).replace(openPattern, escOpen).replace(closePattern, escClose).replace(commaPattern, escComma).replace(periodPattern, escPeriod); } function unescapeBraces(str2) { - return str2.split(escSlash).join("\\").split(escOpen).join("{").split(escClose).join("}").split(escComma).join(",").split(escPeriod).join("."); + return str2.replace(escSlashPattern, "\\").replace(escOpenPattern, "{").replace(escClosePattern, "}").replace(escCommaPattern, ",").replace(escPeriodPattern, "."); } function parseCommaParts(str2) { - if (!str2) + if (!str2) { return [""]; - var parts = []; - var m = balanced("{", "}", str2); - if (!m) + } + const parts = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) { return str2.split(","); - var pre = m.pre; - var body = m.body; - var post = m.post; - var p = pre.split(","); + } + const { pre, body, post } = m; + const p = pre.split(","); p[p.length - 1] += "{" + body + "}"; - var postParts = parseCommaParts(post); + const postParts = parseCommaParts(post); if (post.length) { + ; p[p.length - 1] += postParts.shift(); p.push.apply(p, postParts); } parts.push.apply(parts, p); return parts; } - function expandTop(str2) { - if (!str2) + function expand(str2) { + if (!str2) { return []; - if (str2.substr(0, 2) === "{}") { - str2 = "\\{\\}" + str2.substr(2); } - return expand(escapeBraces(str2), true).map(unescapeBraces); + if (str2.slice(0, 2) === "{}") { + str2 = "\\{\\}" + str2.slice(2); + } + return expand_(escapeBraces(str2), true).map(unescapeBraces); } function embrace(str2) { return "{" + str2 + "}"; @@ -94191,73 +94266,74 @@ var require_brace_expansion3 = __commonJS({ function gte5(i, y) { return i >= y; } - function expand(str2, isTop) { - var expansions = []; - var m = balanced("{", "}", str2); - if (!m) return [str2]; - var pre = m.pre; - var post = m.post.length ? expand(m.post, false) : [""]; + function expand_(str2, isTop) { + const expansions = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) + return [str2]; + const pre = m.pre; + const post = m.post.length ? expand_(m.post, false) : [""]; if (/\$$/.test(m.pre)) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + "{" + m.body + "}" + post[k]; + for (let k = 0; k < post.length; k++) { + const expansion = pre + "{" + m.body + "}" + post[k]; expansions.push(expansion); } } else { - var isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); - var isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); - var isSequence = isNumericSequence || isAlphaSequence; - var isOptions = m.body.indexOf(",") >= 0; + const isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); + const isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); + const isSequence = isNumericSequence || isAlphaSequence; + const isOptions = m.body.indexOf(",") >= 0; if (!isSequence && !isOptions) { if (m.post.match(/,(?!,).*\}/)) { str2 = m.pre + "{" + m.body + escClose + m.post; - return expand(str2); + return expand_(str2); } return [str2]; } - var n; + let n; if (isSequence) { n = m.body.split(/\.\./); } else { n = parseCommaParts(m.body); - if (n.length === 1) { - n = expand(n[0], false).map(embrace); + if (n.length === 1 && n[0] !== void 0) { + n = expand_(n[0], false).map(embrace); if (n.length === 1) { - return post.map(function(p) { - return m.pre + n[0] + p; - }); + return post.map((p) => m.pre + n[0] + p); } } } - var N; - if (isSequence) { - var x = numeric(n[0]); - var y = numeric(n[1]); - var width = Math.max(n[0].length, n[1].length); - var incr = n.length == 3 ? Math.abs(numeric(n[2])) : 1; - var test = lte; - var reverse = y < x; + let N; + if (isSequence && n[0] !== void 0 && n[1] !== void 0) { + const x = numeric(n[0]); + const y = numeric(n[1]); + const width = Math.max(n[0].length, n[1].length); + let incr = n.length === 3 && n[2] !== void 0 ? Math.abs(numeric(n[2])) : 1; + let test = lte; + const reverse = y < x; if (reverse) { incr *= -1; test = gte5; } - var pad = n.some(isPadded); + const pad = n.some(isPadded); N = []; - for (var i = x; test(i, y); i += incr) { - var c; + for (let i = x; test(i, y); i += incr) { + let c; if (isAlphaSequence) { c = String.fromCharCode(i); - if (c === "\\") + if (c === "\\") { c = ""; + } } else { c = String(i); if (pad) { - var need = width - c.length; + const need = width - c.length; if (need > 0) { - var z = new Array(need + 1).join("0"); - if (i < 0) + const z = new Array(need + 1).join("0"); + if (i < 0) { c = "-" + z + c.slice(1); - else + } else { c = z + c; + } } } } @@ -94265,15 +94341,16 @@ var require_brace_expansion3 = __commonJS({ } } else { N = []; - for (var j = 0; j < n.length; j++) { - N.push.apply(N, expand(n[j], false)); + for (let j = 0; j < n.length; j++) { + N.push.apply(N, expand_(n[j], false)); } } - for (var j = 0; j < N.length; j++) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + N[j] + post[k]; - if (!isTop || isSequence || expansion) + for (let j = 0; j < N.length; j++) { + for (let k = 0; k < post.length; k++) { + const expansion = pre + N[j] + post[k]; + if (!isTop || isSequence || expansion) { expansions.push(expansion); + } } } } @@ -94282,9 +94359,9 @@ var require_brace_expansion3 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js var require_assert_valid_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.assertValidPattern = void 0; @@ -94301,9 +94378,9 @@ var require_assert_valid_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js var require_brace_expressions = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.parseClass = void 0; @@ -94418,22 +94495,25 @@ var require_brace_expressions = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js var require_unescape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = void 0; - var unescape = (s, { windowsPathsNoEscape = false } = {}) => { - return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + var unescape = (s, { windowsPathsNoEscape = false, magicalBraces = true } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + } + return windowsPathsNoEscape ? s.replace(/\[([^\/\\{}])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\{}])\]/g, "$1$2").replace(/\\([^\/{}])/g, "$1"); }; exports2.unescape = unescape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js var require_ast = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.AST = void 0; @@ -94789,7 +94869,7 @@ var require_ast = __commonJS({ if (this.#root === this) this.#fillNegs(); if (!this.type) { - const noEmpty = this.isStart() && this.isEnd(); + const noEmpty = this.isStart() && this.isEnd() && !this.#parts.some((s) => typeof s !== "string"); const src = this.#parts.map((p) => { const [re, _2, hasMagic, uflag] = typeof p === "string" ? _AST.#parseGlob(p, this.#hasMagic, noEmpty) : p.toRegExpSource(allowDot); this.#hasMagic = this.#hasMagic || hasMagic; @@ -94899,10 +94979,7 @@ var require_ast = __commonJS({ } } if (c === "*") { - if (noEmpty && glob2 === "*") - re += starNoEmpty; - else - re += star; + re += noEmpty && glob2 === "*" ? starNoEmpty : star; hasMagic = true; continue; } @@ -94920,29 +94997,29 @@ var require_ast = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js var require_escape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.escape = void 0; - var escape = (s, { windowsPathsNoEscape = false } = {}) => { + var escape = (s, { windowsPathsNoEscape = false, magicalBraces = false } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/[?*()[\]{}]/g, "[$&]") : s.replace(/[?*()[\]\\{}]/g, "\\$&"); + } return windowsPathsNoEscape ? s.replace(/[?*()[\]]/g, "[$&]") : s.replace(/[?*()[\]\\]/g, "\\$&"); }; exports2.escape = escape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js -var require_commonjs13 = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js"(exports2) { +// node_modules/glob/node_modules/minimatch/dist/commonjs/index.js +var require_commonjs15 = __commonJS({ + "node_modules/glob/node_modules/minimatch/dist/commonjs/index.js"(exports2) { "use strict"; - var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { - return mod && mod.__esModule ? mod : { "default": mod }; - }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = exports2.escape = exports2.AST = exports2.Minimatch = exports2.match = exports2.makeRe = exports2.braceExpand = exports2.defaults = exports2.filter = exports2.GLOBSTAR = exports2.sep = exports2.minimatch = void 0; - var brace_expansion_1 = __importDefault4(require_brace_expansion3()); + var brace_expansion_1 = require_commonjs14(); var assert_valid_pattern_js_1 = require_assert_valid_pattern(); var ast_js_1 = require_ast(); var escape_js_1 = require_escape(); @@ -95065,7 +95142,7 @@ var require_commonjs13 = __commonJS({ if (options.nobrace || !/\{(?:(?!\{).)*\}/.test(pattern)) { return [pattern]; } - return (0, brace_expansion_1.default)(pattern); + return (0, brace_expansion_1.expand)(pattern); }; exports2.braceExpand = braceExpand; exports2.minimatch.braceExpand = exports2.braceExpand; @@ -95589,16 +95666,27 @@ var require_commonjs13 = __commonJS({ pp[i] = twoStar; } } else if (next === void 0) { - pp[i - 1] = prev + "(?:\\/|" + twoStar + ")?"; + pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + ")?"; } else if (next !== exports2.GLOBSTAR) { pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + "\\/)" + next; pp[i + 1] = exports2.GLOBSTAR; } }); - return pp.filter((p) => p !== exports2.GLOBSTAR).join("/"); + const filtered = pp.filter((p) => p !== exports2.GLOBSTAR); + if (this.partial && filtered.length >= 1) { + const prefixes = []; + for (let i = 1; i <= filtered.length; i++) { + prefixes.push(filtered.slice(0, i).join("/")); + } + return "(?:" + prefixes.join("|") + ")"; + } + return filtered.join("/"); }).join("|"); const [open, close] = set2.length > 1 ? ["(?:", ")"] : ["", ""]; re = "^" + open + re + close + "$"; + if (this.partial) { + re = "^(?:\\/|" + open + re.slice(1, -1) + close + ")$"; + } if (this.negate) re = "^(?!" + re + ").+$"; try { @@ -95685,9 +95773,9 @@ var require_commonjs13 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js -var require_commonjs14 = __commonJS({ - "node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js"(exports2) { +// node_modules/lru-cache/dist/commonjs/index.js +var require_commonjs16 = __commonJS({ + "node_modules/lru-cache/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.LRUCache = void 0; @@ -95776,6 +95864,7 @@ var require_commonjs14 = __commonJS({ #max; #maxSize; #dispose; + #onInsert; #disposeAfter; #fetchMethod; #memoMethod; @@ -95857,6 +95946,7 @@ var require_commonjs14 = __commonJS({ #hasDispose; #hasFetchMethod; #hasDisposeAfter; + #hasOnInsert; /** * Do not call this method unless you need to inspect the * inner workings of the cache. If anything returned by this @@ -95933,6 +96023,12 @@ var require_commonjs14 = __commonJS({ get dispose() { return this.#dispose; } + /** + * {@link LRUCache.OptionsBase.onInsert} (read-only) + */ + get onInsert() { + return this.#onInsert; + } /** * {@link LRUCache.OptionsBase.disposeAfter} (read-only) */ @@ -95940,7 +96036,7 @@ var require_commonjs14 = __commonJS({ return this.#disposeAfter; } constructor(options) { - const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; + const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, onInsert, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; if (max !== 0 && !isPosInt(max)) { throw new TypeError("max option must be a nonnegative integer"); } @@ -95982,6 +96078,9 @@ var require_commonjs14 = __commonJS({ if (typeof dispose === "function") { this.#dispose = dispose; } + if (typeof onInsert === "function") { + this.#onInsert = onInsert; + } if (typeof disposeAfter === "function") { this.#disposeAfter = disposeAfter; this.#disposed = []; @@ -95990,6 +96089,7 @@ var require_commonjs14 = __commonJS({ this.#disposed = void 0; } this.#hasDispose = !!this.#dispose; + this.#hasOnInsert = !!this.#onInsert; this.#hasDisposeAfter = !!this.#disposeAfter; this.noDisposeOnSet = !!noDisposeOnSet; this.noUpdateTTL = !!noUpdateTTL; @@ -96392,7 +96492,7 @@ var require_commonjs14 = __commonJS({ } /** * Return an array of [key, {@link LRUCache.Entry}] tuples which can be - * passed to {@link LRLUCache#load}. + * passed to {@link LRUCache#load}. * * The `start` fields are calculated relative to a portable `Date.now()` * timestamp, even if `performance.now()` is available. @@ -96503,6 +96603,9 @@ var require_commonjs14 = __commonJS({ if (status) status.set = "add"; noUpdateTTL = false; + if (this.#hasOnInsert) { + this.#onInsert?.(v, k, "add"); + } } else { this.#moveToTail(index); const oldVal = this.#valList[index]; @@ -96538,6 +96641,9 @@ var require_commonjs14 = __commonJS({ } else if (status) { status.set = "update"; } + if (this.#hasOnInsert) { + this.onInsert?.(v, k, v === oldVal ? "update" : "replace"); + } } if (ttl !== 0 && !this.#ttls) { this.#initializeTTLTracking(); @@ -97063,7 +97169,7 @@ var require_commonjs14 = __commonJS({ }); // node_modules/minipass/dist/commonjs/index.js -var require_commonjs15 = __commonJS({ +var require_commonjs17 = __commonJS({ "node_modules/minipass/dist/commonjs/index.js"(exports2) { "use strict"; var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { @@ -97954,9 +98060,9 @@ var require_commonjs15 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js -var require_commonjs16 = __commonJS({ - "node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js"(exports2) { +// node_modules/path-scurry/dist/commonjs/index.js +var require_commonjs18 = __commonJS({ + "node_modules/path-scurry/dist/commonjs/index.js"(exports2) { "use strict"; var __createBinding4 = exports2 && exports2.__createBinding || (Object.create ? (function(o, m, k, k2) { if (k2 === void 0) k2 = k; @@ -97987,14 +98093,14 @@ var require_commonjs16 = __commonJS({ }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.PathScurry = exports2.Path = exports2.PathScurryDarwin = exports2.PathScurryPosix = exports2.PathScurryWin32 = exports2.PathScurryBase = exports2.PathPosix = exports2.PathWin32 = exports2.PathBase = exports2.ChildrenCache = exports2.ResolveCache = void 0; - var lru_cache_1 = require_commonjs14(); + var lru_cache_1 = require_commonjs16(); var node_path_1 = require("node:path"); var node_url_1 = require("node:url"); var fs_1 = require("fs"); var actualFS = __importStar4(require("node:fs")); var realpathSync = fs_1.realpathSync.native; var promises_1 = require("node:fs/promises"); - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var defaultFS = { lstatSync: fs_1.lstatSync, readdir: fs_1.readdir, @@ -98209,6 +98315,8 @@ var require_commonjs16 = __commonJS({ /** * Deprecated alias for Dirent['parentPath'] Somewhat counterintuitively, * this property refers to the *parent* path, not the path object itself. + * + * @deprecated */ get path() { return this.parentPath; @@ -99728,13 +99836,13 @@ var require_commonjs16 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js +// node_modules/glob/dist/commonjs/pattern.js var require_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js"(exports2) { + "node_modules/glob/dist/commonjs/pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Pattern = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var isPatternList = (pl) => pl.length >= 1; var isGlobList = (gl) => gl.length >= 1; var Pattern = class _Pattern { @@ -99902,13 +100010,13 @@ var require_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js +// node_modules/glob/dist/commonjs/ignore.js var require_ignore = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js"(exports2) { + "node_modules/glob/dist/commonjs/ignore.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Ignore = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var pattern_js_1 = require_pattern(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; var Ignore = class { @@ -99999,13 +100107,13 @@ var require_ignore = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js +// node_modules/glob/dist/commonjs/processor.js var require_processor = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js"(exports2) { + "node_modules/glob/dist/commonjs/processor.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Processor = exports2.SubWalks = exports2.MatchRecord = exports2.HasWalkedCache = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var HasWalkedCache = class _HasWalkedCache { store; constructor(store = /* @__PURE__ */ new Map()) { @@ -100232,13 +100340,13 @@ var require_processor = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js +// node_modules/glob/dist/commonjs/walker.js var require_walker = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js"(exports2) { + "node_modules/glob/dist/commonjs/walker.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.GlobStream = exports2.GlobWalker = exports2.GlobUtil = void 0; - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var ignore_js_1 = require_ignore(); var processor_js_1 = require_processor(); var makeIgnore = (ignore, opts) => typeof ignore === "string" ? new ignore_js_1.Ignore([ignore], opts) : Array.isArray(ignore) ? new ignore_js_1.Ignore(ignore, opts) : ignore; @@ -100572,15 +100680,15 @@ var require_walker = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js +// node_modules/glob/dist/commonjs/glob.js var require_glob2 = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js"(exports2) { + "node_modules/glob/dist/commonjs/glob.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Glob = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var node_url_1 = require("node:url"); - var path_scurry_1 = require_commonjs16(); + var path_scurry_1 = require_commonjs18(); var pattern_js_1 = require_pattern(); var walker_js_1 = require_walker(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; @@ -100785,13 +100893,13 @@ var require_glob2 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js +// node_modules/glob/dist/commonjs/has-magic.js var require_has_magic = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js"(exports2) { + "node_modules/glob/dist/commonjs/has-magic.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.hasMagic = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var hasMagic = (pattern, options = {}) => { if (!Array.isArray(pattern)) { pattern = [pattern]; @@ -100806,9 +100914,9 @@ var require_has_magic = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js -var require_commonjs17 = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js"(exports2) { +// node_modules/glob/dist/commonjs/index.js +var require_commonjs19 = __commonJS({ + "node_modules/glob/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.glob = exports2.sync = exports2.iterate = exports2.iterateSync = exports2.stream = exports2.streamSync = exports2.Ignore = exports2.hasMagic = exports2.Glob = exports2.unescape = exports2.escape = void 0; @@ -100817,10 +100925,10 @@ var require_commonjs17 = __commonJS({ exports2.globSync = globSync; exports2.globIterateSync = globIterateSync; exports2.globIterate = globIterate; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var glob_js_1 = require_glob2(); var has_magic_js_1 = require_has_magic(); - var minimatch_2 = require_commonjs13(); + var minimatch_2 = require_commonjs15(); Object.defineProperty(exports2, "escape", { enumerable: true, get: function() { return minimatch_2.escape; } }); @@ -100897,7 +101005,7 @@ var require_file3 = __commonJS({ var difference = require_difference(); var union = require_union(); var isPlainObject = require_isPlainObject(); - var glob2 = require_commonjs17(); + var glob2 = require_commonjs19(); var file = module2.exports = {}; var pathSeparatorRe = /[\/\\]/g; var processPatterns = function(patterns, fn) { @@ -102624,18 +102732,18 @@ var require_zip_archive_output_stream = __commonJS({ ZipArchiveOutputStream.prototype._smartStream = function(ae, callback) { var deflate = ae.getMethod() === constants.METHOD_DEFLATED; var process2 = deflate ? new DeflateCRC32Stream(this.options.zlib) : new CRC32Stream(); - var error4 = null; + var error3 = null; function handleStuff() { var digest = process2.digest().readUInt32BE(0); ae.setCrc(digest); ae.setSize(process2.size()); ae.setCompressedSize(process2.size(true)); this._afterAppend(ae); - callback(error4, ae); + callback(error3, ae); } process2.once("end", handleStuff.bind(this)); process2.once("error", function(err) { - error4 = err; + error3 = err; }); process2.pipe(this, { end: false }); return process2; @@ -104001,11 +104109,11 @@ var require_streamx = __commonJS({ } [asyncIterator]() { const stream = this; - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("readable", onreadable); this.on("close", onclose); @@ -104037,7 +104145,7 @@ var require_streamx = __commonJS({ } function ondata(data) { if (promiseReject === null) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else if (data === null && (stream._duplexState & READ_DONE) === 0) promiseReject(STREAM_DESTROYED); else promiseResolve({ value: data, done: data === null }); promiseReject = promiseResolve = null; @@ -104211,7 +104319,7 @@ var require_streamx = __commonJS({ if (all.length < 2) throw new Error("Pipeline requires at least 2 streams"); let src = all[0]; let dest = null; - let error4 = null; + let error3 = null; for (let i = 1; i < all.length; i++) { dest = all[i]; if (isStreamx(src)) { @@ -104226,14 +104334,14 @@ var require_streamx = __commonJS({ let fin = false; const autoDestroy = isStreamx(dest) || !!(dest._writableState && dest._writableState.autoDestroy); dest.on("error", (err) => { - if (error4 === null) error4 = err; + if (error3 === null) error3 = err; }); dest.on("finish", () => { fin = true; - if (!autoDestroy) done(error4); + if (!autoDestroy) done(error3); }); if (autoDestroy) { - dest.on("close", () => done(error4 || (fin ? null : PREMATURE_CLOSE))); + dest.on("close", () => done(error3 || (fin ? null : PREMATURE_CLOSE))); } } return dest; @@ -104246,8 +104354,8 @@ var require_streamx = __commonJS({ } } function onerror(err) { - if (!err || error4) return; - error4 = err; + if (!err || error3) return; + error3 = err; for (const s of all) { s.destroy(err); } @@ -104813,7 +104921,7 @@ var require_extract = __commonJS({ cb(null); } [Symbol.asyncIterator]() { - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; let entryStream = null; @@ -104821,7 +104929,7 @@ var require_extract = __commonJS({ const extract2 = this; this.on("entry", onentry); this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("close", onclose); return { @@ -104845,8 +104953,8 @@ var require_extract = __commonJS({ cb(err); } function onnext(resolve2, reject) { - if (error4) { - return reject(error4); + if (error3) { + return reject(error3); } if (entryStream) { resolve2({ value: entryStream, done: false }); @@ -104872,9 +104980,9 @@ var require_extract = __commonJS({ } } function onclose() { - consumeCallback(error4); + consumeCallback(error3); if (!promiseResolve) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else promiseResolve({ value: void 0, done: true }); promiseResolve = promiseReject = null; } @@ -105770,18 +105878,18 @@ var require_zip2 = __commonJS({ return zipUploadStream; }); } - var zipErrorCallback = (error4) => { + var zipErrorCallback = (error3) => { core14.error("An error has occurred while creating the zip file for upload"); - core14.info(error4); + core14.info(error3); throw new Error("An error has occurred during zip creation for the artifact"); }; - var zipWarningCallback = (error4) => { - if (error4.code === "ENOENT") { + var zipWarningCallback = (error3) => { + if (error3.code === "ENOENT") { core14.warning("ENOENT warning during artifact zip creation. No such file or directory"); - core14.info(error4); + core14.info(error3); } else { - core14.warning(`A non-blocking warning has occurred during artifact zip creation: ${error4.code}`); - core14.info(error4); + core14.warning(`A non-blocking warning has occurred during artifact zip creation: ${error3.code}`); + core14.info(error3); } }; var zipFinishCallback = () => { @@ -107588,8 +107696,8 @@ var require_parser_stream = __commonJS({ this.unzipStream.on("entry", function(entry) { self2.push(entry); }); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(ParserStream, Transform); @@ -107729,8 +107837,8 @@ var require_extract2 = __commonJS({ this.createdDirectories = {}; var self2 = this; this.unzipStream.on("entry", this._processEntry.bind(this)); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(Extract, Transform); @@ -107764,8 +107872,8 @@ var require_extract2 = __commonJS({ self2.unfinishedEntries--; self2._notifyAwaiter(); }); - pipedStream.on("error", function(error4) { - self2.emit("error", error4); + pipedStream.on("error", function(error3) { + self2.emit("error", error3); }); entry.pipe(pipedStream); }; @@ -107900,11 +108008,11 @@ var require_download_artifact = __commonJS({ try { yield promises_1.default.access(path2); return true; - } catch (error4) { - if (error4.code === "ENOENT") { + } catch (error3) { + if (error3.code === "ENOENT") { return false; } else { - throw error4; + throw error3; } } }); @@ -107915,9 +108023,9 @@ var require_download_artifact = __commonJS({ while (retryCount < 5) { try { return yield streamExtractExternal(url, directory); - } catch (error4) { + } catch (error3) { retryCount++; - core14.debug(`Failed to download artifact after ${retryCount} retries due to ${error4.message}. Retrying in 5 seconds...`); + core14.debug(`Failed to download artifact after ${retryCount} retries due to ${error3.message}. Retrying in 5 seconds...`); yield new Promise((resolve2) => setTimeout(resolve2, 5e3)); } } @@ -107946,10 +108054,10 @@ var require_download_artifact = __commonJS({ const extractStream = passThrough; extractStream.on("data", () => { timer.refresh(); - }).on("error", (error4) => { - core14.debug(`response.message: Artifact download failed: ${error4.message}`); + }).on("error", (error3) => { + core14.debug(`response.message: Artifact download failed: ${error3.message}`); clearTimeout(timer); - reject(error4); + reject(error3); }).pipe(unzip_stream_1.default.Extract({ path: directory })).on("close", () => { clearTimeout(timer); if (hashStream) { @@ -107958,8 +108066,8 @@ var require_download_artifact = __commonJS({ core14.info(`SHA256 digest of downloaded artifact is ${sha256Digest}`); } resolve2({ sha256Digest: `sha256:${sha256Digest}` }); - }).on("error", (error4) => { - reject(error4); + }).on("error", (error3) => { + reject(error3); }); }); }); @@ -107998,8 +108106,8 @@ var require_download_artifact = __commonJS({ core14.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -108041,8 +108149,8 @@ Are you trying to download from a different run? Try specifying a github-token w core14.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -108140,9 +108248,9 @@ var require_dist_node16 = __commonJS({ return request(options).then((response) => { octokit.log.info(`${requestOptions.method} ${path2} - ${response.status} in ${Date.now() - start}ms`); return response; - }).catch((error4) => { - octokit.log.info(`${requestOptions.method} ${path2} - ${error4.status} in ${Date.now() - start}ms`); - throw error4; + }).catch((error3) => { + octokit.log.info(`${requestOptions.method} ${path2} - ${error3.status} in ${Date.now() - start}ms`); + throw error3; }); }); } @@ -108160,22 +108268,22 @@ var require_dist_node17 = __commonJS({ return ex && typeof ex === "object" && "default" in ex ? ex["default"] : ex; } var Bottleneck = _interopDefault(require_light()); - async function errorRequest(octokit, state, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(octokit, state, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } async function wrapRequest(state, request, options) { const limiter = new Bottleneck(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -108197,12 +108305,12 @@ var require_dist_node17 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -108693,13 +108801,13 @@ var require_client2 = __commonJS({ throw new errors_1.GHESNotSupportedError(); } return (0, upload_artifact_1.uploadArtifact)(name, files, rootDirectory, options); - } catch (error4) { - (0, core_1.warning)(`Artifact upload failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Artifact upload failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions is operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -108714,13 +108822,13 @@ If the error persists, please check whether Actions is operating normally at [ht return (0, download_artifact_1.downloadArtifactPublic)(artifactId, repositoryOwner, repositoryName, token, downloadOptions); } return (0, download_artifact_1.downloadArtifactInternal)(artifactId, options); - } catch (error4) { - (0, core_1.warning)(`Download Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Download Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -108735,13 +108843,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, list_artifacts_1.listArtifactsPublic)(workflowRunId, repositoryOwner, repositoryName, token, options === null || options === void 0 ? void 0 : options.latest); } return (0, list_artifacts_1.listArtifactsInternal)(options === null || options === void 0 ? void 0 : options.latest); - } catch (error4) { - (0, core_1.warning)(`Listing Artifacts failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Listing Artifacts failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -108756,13 +108864,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, get_artifact_1.getArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, get_artifact_1.getArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Get Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Get Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -108777,13 +108885,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, delete_artifact_1.deleteArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, delete_artifact_1.deleteArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Delete Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Delete Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -109283,14 +109391,14 @@ var require_tmp = __commonJS({ options.template = _getRelativePathSync("template", options.template, tmpDir); return options; } - function _isEBADF(error4) { - return _isExpectedError(error4, -EBADF, "EBADF"); + function _isEBADF(error3) { + return _isExpectedError(error3, -EBADF, "EBADF"); } - function _isENOENT(error4) { - return _isExpectedError(error4, -ENOENT, "ENOENT"); + function _isENOENT(error3) { + return _isExpectedError(error3, -ENOENT, "ENOENT"); } - function _isExpectedError(error4, errno, code) { - return IS_WIN32 ? error4.code === code : error4.code === code && error4.errno === errno; + function _isExpectedError(error3, errno, code) { + return IS_WIN32 ? error3.code === code : error3.code === code && error3.errno === errno; } function setGracefulCleanup() { _gracefulCleanup = true; @@ -111002,9 +111110,9 @@ var require_upload_gzip = __commonJS({ const size = (yield stat(tempFilePath)).size; resolve2(size); })); - outputStream.on("error", (error4) => { - console.log(error4); - reject(error4); + outputStream.on("error", (error3) => { + console.log(error3); + reject(error3); }); }); }); @@ -111129,9 +111237,9 @@ var require_requestUtils2 = __commonJS({ } isRetryable = (0, utils_1.isRetryableStatusCode)(statusCode); errorMessage = `Artifact service responded with ${statusCode}`; - } catch (error4) { + } catch (error3) { isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { core14.info(`${name} - Error is not retryable`); @@ -111497,9 +111605,9 @@ var require_upload_http_client = __commonJS({ let response; try { response = yield uploadChunkRequest(); - } catch (error4) { + } catch (error3) { core14.info(`An error has been caught http-client index ${httpClientIndex}, retrying the upload`); - console.log(error4); + console.log(error3); if (incrementAndCheckRetryLimit()) { return false; } @@ -111688,8 +111796,8 @@ var require_download_http_client = __commonJS({ } this.statusReporter.incrementProcessedCount(); } - }))).catch((error4) => { - throw new Error(`Unable to download the artifact: ${error4}`); + }))).catch((error3) => { + throw new Error(`Unable to download the artifact: ${error3}`); }).finally(() => { this.statusReporter.stop(); this.downloadHttpManager.disposeAndReplaceAllClients(); @@ -111754,9 +111862,9 @@ var require_download_http_client = __commonJS({ let response; try { response = yield makeDownloadRequest(); - } catch (error4) { + } catch (error3) { core14.info("An error occurred while attempting to download a file"); - console.log(error4); + console.log(error3); yield backOff(); continue; } @@ -111770,7 +111878,7 @@ var require_download_http_client = __commonJS({ } else { forceRetry = true; } - } catch (error4) { + } catch (error3) { forceRetry = true; } } @@ -111796,31 +111904,31 @@ var require_download_http_client = __commonJS({ yield new Promise((resolve2, reject) => { if (isGzip) { const gunzip = zlib.createGunzip(); - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core14.info(`An error occurred while attempting to read the response stream`); gunzip.close(); destinationStream.close(); - reject(error4); - }).pipe(gunzip).on("error", (error4) => { + reject(error3); + }).pipe(gunzip).on("error", (error3) => { core14.info(`An error occurred while attempting to decompress the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve2(); - }).on("error", (error4) => { + }).on("error", (error3) => { core14.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } else { - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core14.info(`An error occurred while attempting to read the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve2(); - }).on("error", (error4) => { + }).on("error", (error3) => { core14.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } }); @@ -114827,7 +114935,7 @@ var require_debug2 = __commonJS({ if (!debug4) { try { debug4 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug4 !== "function") { debug4 = function() { @@ -114860,8 +114968,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -114935,9 +115043,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -115104,10 +115212,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -115306,12 +115414,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -116443,7 +116551,6 @@ var io2 = __toESM(require_io2()); // src/util.ts var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js @@ -116461,14 +116568,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs2.lstat(itemPath, { bigint: true }) : await fs2.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs2.lstat(itemPath, { bigint: true }) : await fs2.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs2.readdir(itemPath) : await fs2.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs2.readdir(itemPath) : await fs2.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -116479,13 +116586,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -117365,6 +117472,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -117484,7 +117603,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -117524,16 +117643,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -119167,8 +119277,8 @@ var ConfigurationError = class extends Error { super(message); } }; -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } // src/actions-util.ts @@ -119304,7 +119414,7 @@ function getActionsLogger() { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; @@ -119372,6 +119482,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -119483,21 +119598,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, @@ -119510,6 +119625,8 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, @@ -119789,9 +119906,9 @@ async function runWrapper() { } ); } - } catch (error4) { + } catch (error3) { logger.warning( - `start-proxy post-action step failed: ${getErrorMessage(error4)}` + `start-proxy post-action step failed: ${getErrorMessage(error3)}` ); } } @@ -119880,5 +119997,5 @@ tmp/lib/tmp.js: *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/start-proxy-action.js b/lib/start-proxy-action.js index 5613603cb..b8bed55ea 100644 --- a/lib/start-proxy-action.js +++ b/lib/start-proxy-action.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug5("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve2(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug5; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning6(message, properties = {}) { + exports2.error = error3; + function warning7(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning6; + exports2.warning = warning7; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -24795,10 +24795,10 @@ var require_util8 = __commonJS({ rval = api.setItem(id, obj); } if (typeof rval !== "undefined" && rval.rval !== true) { - var error4 = new Error(rval.error.message); - error4.id = rval.error.id; - error4.name = rval.error.name; - throw error4; + var error3 = new Error(rval.error.message); + error3.id = rval.error.id; + error3.name = rval.error.name; + throw error3; } }; var _getStorageObject = function(api, id) { @@ -24809,10 +24809,10 @@ var require_util8 = __commonJS({ if (api.init) { if (rval.rval === null) { if (rval.error) { - var error4 = new Error(rval.error.message); - error4.id = rval.error.id; - error4.name = rval.error.name; - throw error4; + var error3 = new Error(rval.error.message); + error3.id = rval.error.id; + error3.name = rval.error.name; + throw error3; } rval = null; } else { @@ -26476,11 +26476,11 @@ var require_asn1 = __commonJS({ }; function _checkBufferLength(bytes, remaining, n) { if (n > remaining) { - var error4 = new Error("Too few bytes to parse DER."); - error4.available = bytes.length(); - error4.remaining = remaining; - error4.requested = n; - throw error4; + var error3 = new Error("Too few bytes to parse DER."); + error3.available = bytes.length(); + error3.remaining = remaining; + error3.requested = n; + throw error3; } } var _getValueLength = function(bytes, remaining) { @@ -26533,10 +26533,10 @@ var require_asn1 = __commonJS({ var byteCount = bytes.length(); var value = _fromDer(bytes, bytes.length(), 0, options); if (options.parseAllBytes && bytes.length() !== 0) { - var error4 = new Error("Unparsed DER bytes remain after ASN.1 parsing."); - error4.byteCount = byteCount; - error4.remaining = bytes.length(); - throw error4; + var error3 = new Error("Unparsed DER bytes remain after ASN.1 parsing."); + error3.byteCount = byteCount; + error3.remaining = bytes.length(); + throw error3; } return value; }; @@ -26552,11 +26552,11 @@ var require_asn1 = __commonJS({ remaining -= start - bytes.length(); if (length !== void 0 && length > remaining) { if (options.strict) { - var error4 = new Error("Too few bytes to read ASN.1 value."); - error4.available = bytes.length(); - error4.remaining = remaining; - error4.requested = length; - throw error4; + var error3 = new Error("Too few bytes to read ASN.1 value."); + error3.available = bytes.length(); + error3.remaining = remaining; + error3.requested = length; + throw error3; } length = remaining; } @@ -26880,9 +26880,9 @@ var require_asn1 = __commonJS({ if (x >= -2147483648 && x < 2147483648) { return rval.putSignedInt(x, 32); } - var error4 = new Error("Integer too large; max is 32-bits."); - error4.integer = x; - throw error4; + var error3 = new Error("Integer too large; max is 32-bits."); + error3.integer = x; + throw error3; }; asn1.derToInteger = function(bytes) { if (typeof bytes === "string") { @@ -30475,10 +30475,10 @@ var require_pkcs1 = __commonJS({ var keyLength = Math.ceil(key.n.bitLength() / 8); var maxLength = keyLength - 2 * md.digestLength - 2; if (message.length > maxLength) { - var error4 = new Error("RSAES-OAEP input message length is too long."); - error4.length = message.length; - error4.maxLength = maxLength; - throw error4; + var error3 = new Error("RSAES-OAEP input message length is too long."); + error3.length = message.length; + error3.maxLength = maxLength; + throw error3; } if (!label) { label = ""; @@ -30494,10 +30494,10 @@ var require_pkcs1 = __commonJS({ if (!seed) { seed = forge.random.getBytes(md.digestLength); } else if (seed.length !== md.digestLength) { - var error4 = new Error("Invalid RSAES-OAEP seed. The seed length must match the digest length."); - error4.seedLength = seed.length; - error4.digestLength = md.digestLength; - throw error4; + var error3 = new Error("Invalid RSAES-OAEP seed. The seed length must match the digest length."); + error3.seedLength = seed.length; + error3.digestLength = md.digestLength; + throw error3; } var dbMask = rsa_mgf1(seed, keyLength - md.digestLength - 1, mgf1Md); var maskedDB = forge.util.xorBytes(DB, dbMask, DB.length); @@ -30521,10 +30521,10 @@ var require_pkcs1 = __commonJS({ } var keyLength = Math.ceil(key.n.bitLength() / 8); if (em.length !== keyLength) { - var error4 = new Error("RSAES-OAEP encoded message length is invalid."); - error4.length = em.length; - error4.expectedLength = keyLength; - throw error4; + var error3 = new Error("RSAES-OAEP encoded message length is invalid."); + error3.length = em.length; + error3.expectedLength = keyLength; + throw error3; } if (md === void 0) { md = forge.md.sha1.create(); @@ -30550,9 +30550,9 @@ var require_pkcs1 = __commonJS({ var dbMask = rsa_mgf1(seed, keyLength - md.digestLength - 1, mgf1Md); var db = forge.util.xorBytes(maskedDB, dbMask, maskedDB.length); var lHashPrime = db.substring(0, md.digestLength); - var error4 = y !== "\0"; + var error3 = y !== "\0"; for (var i = 0; i < md.digestLength; ++i) { - error4 |= lHash.charAt(i) !== lHashPrime.charAt(i); + error3 |= lHash.charAt(i) !== lHashPrime.charAt(i); } var in_ps = 1; var index = md.digestLength; @@ -30560,11 +30560,11 @@ var require_pkcs1 = __commonJS({ var code = db.charCodeAt(j); var is_0 = code & 1 ^ 1; var error_mask = in_ps ? 65534 : 0; - error4 |= code & error_mask; + error3 |= code & error_mask; in_ps = in_ps & is_0; index += in_ps; } - if (error4 || db.charCodeAt(index) !== 1) { + if (error3 || db.charCodeAt(index) !== 1) { throw new Error("Invalid RSAES-OAEP padding."); } return db.substring(index + 1); @@ -30978,9 +30978,9 @@ var require_rsa = __commonJS({ if (md.algorithm in pki2.oids) { oid = pki2.oids[md.algorithm]; } else { - var error4 = new Error("Unknown message digest algorithm."); - error4.algorithm = md.algorithm; - throw error4; + var error3 = new Error("Unknown message digest algorithm."); + error3.algorithm = md.algorithm; + throw error3; } var oidBytes = asn1.oidToDer(oid).getBytes(); var digestInfo = asn1.create( @@ -31076,10 +31076,10 @@ var require_rsa = __commonJS({ pki2.rsa.decrypt = function(ed, key, pub, ml) { var k = Math.ceil(key.n.bitLength() / 8); if (ed.length !== k) { - var error4 = new Error("Encrypted message length is invalid."); - error4.length = ed.length; - error4.expected = k; - throw error4; + var error3 = new Error("Encrypted message length is invalid."); + error3.length = ed.length; + error3.expected = k; + throw error3; } var y = new BigInteger(forge.util.createBuffer(ed).toHex(), 16); if (y.compareTo(key.n) >= 0) { @@ -31451,19 +31451,19 @@ var require_rsa = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, digestInfoValidator, capture, errors)) { - var error4 = new Error( + var error3 = new Error( "ASN.1 object does not contain a valid RSASSA-PKCS1-v1_5 DigestInfo value." ); - error4.errors = errors; - throw error4; + error3.errors = errors; + throw error3; } var oid = asn1.derToOid(capture.algorithmIdentifier); if (!(oid === forge.oids.md2 || oid === forge.oids.md5 || oid === forge.oids.sha1 || oid === forge.oids.sha224 || oid === forge.oids.sha256 || oid === forge.oids.sha384 || oid === forge.oids.sha512 || oid === forge.oids["sha512-224"] || oid === forge.oids["sha512-256"])) { - var error4 = new Error( + var error3 = new Error( "Unknown RSASSA-PKCS1-v1_5 DigestAlgorithm identifier." ); - error4.oid = oid; - throw error4; + error3.oid = oid; + throw error3; } if (oid === forge.oids.md2 || oid === forge.oids.md5) { if (!("parameters" in capture)) { @@ -31579,9 +31579,9 @@ var require_rsa = __commonJS({ capture = {}; errors = []; if (!asn1.validate(obj, rsaPrivateKeyValidator, capture, errors)) { - var error4 = new Error("Cannot read private key. ASN.1 object does not contain an RSAPrivateKey."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read private key. ASN.1 object does not contain an RSAPrivateKey."); + error3.errors = errors; + throw error3; } var n, e, d, p, q, dP, dQ, qInv; n = forge.util.createBuffer(capture.privateKeyModulus).toHex(); @@ -31676,17 +31676,17 @@ var require_rsa = __commonJS({ if (asn1.validate(obj, publicKeyValidator, capture, errors)) { var oid = asn1.derToOid(capture.publicKeyOid); if (oid !== pki2.oids.rsaEncryption) { - var error4 = new Error("Cannot read public key. Unknown OID."); - error4.oid = oid; - throw error4; + var error3 = new Error("Cannot read public key. Unknown OID."); + error3.oid = oid; + throw error3; } obj = capture.rsaPublicKey; } errors = []; if (!asn1.validate(obj, rsaPublicKeyValidator, capture, errors)) { - var error4 = new Error("Cannot read public key. ASN.1 object does not contain an RSAPublicKey."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read public key. ASN.1 object does not contain an RSAPublicKey."); + error3.errors = errors; + throw error3; } var n = forge.util.createBuffer(capture.publicKeyModulus).toHex(); var e = forge.util.createBuffer(capture.publicKeyExponent).toHex(); @@ -31737,10 +31737,10 @@ var require_rsa = __commonJS({ var eb = forge.util.createBuffer(); var k = Math.ceil(key.n.bitLength() / 8); if (m.length > k - 11) { - var error4 = new Error("Message is too long for PKCS#1 v1.5 padding."); - error4.length = m.length; - error4.max = k - 11; - throw error4; + var error3 = new Error("Message is too long for PKCS#1 v1.5 padding."); + error3.length = m.length; + error3.max = k - 11; + throw error3; } eb.putByte(0); eb.putByte(bt); @@ -32134,9 +32134,9 @@ var require_pbe = __commonJS({ cipherFn = forge.des.createEncryptionCipher; break; default: - var error4 = new Error("Cannot encrypt private key. Unknown encryption algorithm."); - error4.algorithm = options.algorithm; - throw error4; + var error3 = new Error("Cannot encrypt private key. Unknown encryption algorithm."); + error3.algorithm = options.algorithm; + throw error3; } var prfAlgorithm = "hmacWith" + options.prfAlgorithm.toUpperCase(); var md = prfAlgorithmToMessageDigest(prfAlgorithm); @@ -32226,9 +32226,9 @@ var require_pbe = __commonJS({ ] ); } else { - var error4 = new Error("Cannot encrypt private key. Unknown encryption algorithm."); - error4.algorithm = options.algorithm; - throw error4; + var error3 = new Error("Cannot encrypt private key. Unknown encryption algorithm."); + error3.algorithm = options.algorithm; + throw error3; } var rval = asn1.create(asn1.Class.UNIVERSAL, asn1.Type.SEQUENCE, true, [ // encryptionAlgorithm @@ -32248,9 +32248,9 @@ var require_pbe = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, encryptedPrivateKeyValidator, capture, errors)) { - var error4 = new Error("Cannot read encrypted private key. ASN.1 object is not a supported EncryptedPrivateKeyInfo."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read encrypted private key. ASN.1 object is not a supported EncryptedPrivateKeyInfo."); + error3.errors = errors; + throw error3; } var oid = asn1.derToOid(capture.encryptionOid); var cipher = pki2.pbe.getCipher(oid, capture.encryptionParams, password); @@ -32271,9 +32271,9 @@ var require_pbe = __commonJS({ pki2.encryptedPrivateKeyFromPem = function(pem) { var msg = forge.pem.decode(pem)[0]; if (msg.type !== "ENCRYPTED PRIVATE KEY") { - var error4 = new Error('Could not convert encrypted private key from PEM; PEM header type is "ENCRYPTED PRIVATE KEY".'); - error4.headerType = msg.type; - throw error4; + var error3 = new Error('Could not convert encrypted private key from PEM; PEM header type is "ENCRYPTED PRIVATE KEY".'); + error3.headerType = msg.type; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { throw new Error("Could not convert encrypted private key from PEM; PEM is encrypted."); @@ -32323,9 +32323,9 @@ var require_pbe = __commonJS({ cipherFn = forge.des.createEncryptionCipher; break; default: - var error4 = new Error('Could not encrypt RSA private key; unsupported encryption algorithm "' + options.algorithm + '".'); - error4.algorithm = options.algorithm; - throw error4; + var error3 = new Error('Could not encrypt RSA private key; unsupported encryption algorithm "' + options.algorithm + '".'); + error3.algorithm = options.algorithm; + throw error3; } var dk = forge.pbe.opensslDeriveBytes(password, iv.substr(0, 8), dkLen); var cipher = cipherFn(dk); @@ -32350,9 +32350,9 @@ var require_pbe = __commonJS({ var rval = null; var msg = forge.pem.decode(pem)[0]; if (msg.type !== "ENCRYPTED PRIVATE KEY" && msg.type !== "PRIVATE KEY" && msg.type !== "RSA PRIVATE KEY") { - var error4 = new Error('Could not convert private key from PEM; PEM header type is not "ENCRYPTED PRIVATE KEY", "PRIVATE KEY", or "RSA PRIVATE KEY".'); - error4.headerType = error4; - throw error4; + var error3 = new Error('Could not convert private key from PEM; PEM header type is not "ENCRYPTED PRIVATE KEY", "PRIVATE KEY", or "RSA PRIVATE KEY".'); + error3.headerType = error3; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { var dkLen; @@ -32397,9 +32397,9 @@ var require_pbe = __commonJS({ }; break; default: - var error4 = new Error('Could not decrypt private key; unsupported encryption algorithm "' + msg.dekInfo.algorithm + '".'); - error4.algorithm = msg.dekInfo.algorithm; - throw error4; + var error3 = new Error('Could not decrypt private key; unsupported encryption algorithm "' + msg.dekInfo.algorithm + '".'); + error3.algorithm = msg.dekInfo.algorithm; + throw error3; } var iv = forge.util.hexToBytes(msg.dekInfo.parameters); var dk = forge.pbe.opensslDeriveBytes(password, iv.substr(0, 8), dkLen); @@ -32498,43 +32498,43 @@ var require_pbe = __commonJS({ case pki2.oids["pbewithSHAAnd40BitRC2-CBC"]: return pki2.pbe.getCipherForPKCS12PBE(oid, params, password); default: - var error4 = new Error("Cannot read encrypted PBE data block. Unsupported OID."); - error4.oid = oid; - error4.supportedOids = [ + var error3 = new Error("Cannot read encrypted PBE data block. Unsupported OID."); + error3.oid = oid; + error3.supportedOids = [ "pkcs5PBES2", "pbeWithSHAAnd3-KeyTripleDES-CBC", "pbewithSHAAnd40BitRC2-CBC" ]; - throw error4; + throw error3; } }; pki2.pbe.getCipherForPBES2 = function(oid, params, password) { var capture = {}; var errors = []; if (!asn1.validate(params, PBES2AlgorithmsValidator, capture, errors)) { - var error4 = new Error("Cannot read password-based-encryption algorithm parameters. ASN.1 object is not a supported EncryptedPrivateKeyInfo."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read password-based-encryption algorithm parameters. ASN.1 object is not a supported EncryptedPrivateKeyInfo."); + error3.errors = errors; + throw error3; } oid = asn1.derToOid(capture.kdfOid); if (oid !== pki2.oids["pkcs5PBKDF2"]) { - var error4 = new Error("Cannot read encrypted private key. Unsupported key derivation function OID."); - error4.oid = oid; - error4.supportedOids = ["pkcs5PBKDF2"]; - throw error4; + var error3 = new Error("Cannot read encrypted private key. Unsupported key derivation function OID."); + error3.oid = oid; + error3.supportedOids = ["pkcs5PBKDF2"]; + throw error3; } oid = asn1.derToOid(capture.encOid); if (oid !== pki2.oids["aes128-CBC"] && oid !== pki2.oids["aes192-CBC"] && oid !== pki2.oids["aes256-CBC"] && oid !== pki2.oids["des-EDE3-CBC"] && oid !== pki2.oids["desCBC"]) { - var error4 = new Error("Cannot read encrypted private key. Unsupported encryption scheme OID."); - error4.oid = oid; - error4.supportedOids = [ + var error3 = new Error("Cannot read encrypted private key. Unsupported encryption scheme OID."); + error3.oid = oid; + error3.supportedOids = [ "aes128-CBC", "aes192-CBC", "aes256-CBC", "des-EDE3-CBC", "desCBC" ]; - throw error4; + throw error3; } var salt = capture.kdfSalt; var count = forge.util.createBuffer(capture.kdfIterationCount); @@ -32574,9 +32574,9 @@ var require_pbe = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(params, pkcs12PbeParamsValidator, capture, errors)) { - var error4 = new Error("Cannot read password-based-encryption algorithm parameters. ASN.1 object is not a supported EncryptedPrivateKeyInfo."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read password-based-encryption algorithm parameters. ASN.1 object is not a supported EncryptedPrivateKeyInfo."); + error3.errors = errors; + throw error3; } var salt = forge.util.createBuffer(capture.salt); var count = forge.util.createBuffer(capture.iterations); @@ -32598,9 +32598,9 @@ var require_pbe = __commonJS({ }; break; default: - var error4 = new Error("Cannot read PKCS #12 PBE data block. Unsupported OID."); - error4.oid = oid; - throw error4; + var error3 = new Error("Cannot read PKCS #12 PBE data block. Unsupported OID."); + error3.oid = oid; + throw error3; } var md = prfOidToMessageDigest(capture.prfOid); var key = pki2.pbe.generatePkcs12Key(password, salt, 1, count, dkLen, md); @@ -32634,16 +32634,16 @@ var require_pbe = __commonJS({ } else { prfAlgorithm = pki2.oids[asn1.derToOid(prfOid)]; if (!prfAlgorithm) { - var error4 = new Error("Unsupported PRF OID."); - error4.oid = prfOid; - error4.supported = [ + var error3 = new Error("Unsupported PRF OID."); + error3.oid = prfOid; + error3.supported = [ "hmacWithSHA1", "hmacWithSHA224", "hmacWithSHA256", "hmacWithSHA384", "hmacWithSHA512" ]; - throw error4; + throw error3; } } return prfAlgorithmToMessageDigest(prfAlgorithm); @@ -32660,16 +32660,16 @@ var require_pbe = __commonJS({ prfAlgorithm = prfAlgorithm.substr(8).toLowerCase(); break; default: - var error4 = new Error("Unsupported PRF algorithm."); - error4.algorithm = prfAlgorithm; - error4.supported = [ + var error3 = new Error("Unsupported PRF algorithm."); + error3.algorithm = prfAlgorithm; + error3.supported = [ "hmacWithSHA1", "hmacWithSHA224", "hmacWithSHA256", "hmacWithSHA384", "hmacWithSHA512" ]; - throw error4; + throw error3; } if (!factory || !(prfAlgorithm in factory)) { throw new Error("Unknown hash algorithm: " + prfAlgorithm); @@ -33666,9 +33666,9 @@ var require_x509 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, rsassaPssParameterValidator, capture, errors)) { - var error4 = new Error("Cannot read RSASSA-PSS parameter block."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read RSASSA-PSS parameter block."); + error3.errors = errors; + throw error3; } if (capture.hashOid !== void 0) { params.hash = params.hash || {}; @@ -33702,11 +33702,11 @@ var require_x509 = __commonJS({ case "RSASSA-PSS": return forge.md.sha256.create(); default: - var error4 = new Error( + var error3 = new Error( "Could not compute " + options.type + " digest. Unknown signature OID." ); - error4.signatureOid = options.signatureOid; - throw error4; + error3.signatureOid = options.signatureOid; + throw error3; } }; var _verifySignature = function(options) { @@ -33721,25 +33721,25 @@ var require_x509 = __commonJS({ var hash, mgf; hash = oids[cert.signatureParameters.mgf.hash.algorithmOid]; if (hash === void 0 || forge.md[hash] === void 0) { - var error4 = new Error("Unsupported MGF hash function."); - error4.oid = cert.signatureParameters.mgf.hash.algorithmOid; - error4.name = hash; - throw error4; + var error3 = new Error("Unsupported MGF hash function."); + error3.oid = cert.signatureParameters.mgf.hash.algorithmOid; + error3.name = hash; + throw error3; } mgf = oids[cert.signatureParameters.mgf.algorithmOid]; if (mgf === void 0 || forge.mgf[mgf] === void 0) { - var error4 = new Error("Unsupported MGF function."); - error4.oid = cert.signatureParameters.mgf.algorithmOid; - error4.name = mgf; - throw error4; + var error3 = new Error("Unsupported MGF function."); + error3.oid = cert.signatureParameters.mgf.algorithmOid; + error3.name = mgf; + throw error3; } mgf = forge.mgf[mgf].create(forge.md[hash].create()); hash = oids[cert.signatureParameters.hash.algorithmOid]; if (hash === void 0 || forge.md[hash] === void 0) { - var error4 = new Error("Unsupported RSASSA-PSS hash function."); - error4.oid = cert.signatureParameters.hash.algorithmOid; - error4.name = hash; - throw error4; + var error3 = new Error("Unsupported RSASSA-PSS hash function."); + error3.oid = cert.signatureParameters.hash.algorithmOid; + error3.name = hash; + throw error3; } scheme = forge.pss.create( forge.md[hash].create(), @@ -33757,11 +33757,11 @@ var require_x509 = __commonJS({ pki2.certificateFromPem = function(pem, computeHash, strict) { var msg = forge.pem.decode(pem)[0]; if (msg.type !== "CERTIFICATE" && msg.type !== "X509 CERTIFICATE" && msg.type !== "TRUSTED CERTIFICATE") { - var error4 = new Error( + var error3 = new Error( 'Could not convert certificate from PEM; PEM header type is not "CERTIFICATE", "X509 CERTIFICATE", or "TRUSTED CERTIFICATE".' ); - error4.headerType = msg.type; - throw error4; + error3.headerType = msg.type; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { throw new Error( @@ -33781,9 +33781,9 @@ var require_x509 = __commonJS({ pki2.publicKeyFromPem = function(pem) { var msg = forge.pem.decode(pem)[0]; if (msg.type !== "PUBLIC KEY" && msg.type !== "RSA PUBLIC KEY") { - var error4 = new Error('Could not convert public key from PEM; PEM header type is not "PUBLIC KEY" or "RSA PUBLIC KEY".'); - error4.headerType = msg.type; - throw error4; + var error3 = new Error('Could not convert public key from PEM; PEM header type is not "PUBLIC KEY" or "RSA PUBLIC KEY".'); + error3.headerType = msg.type; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { throw new Error("Could not convert public key from PEM; PEM is encrypted."); @@ -33839,9 +33839,9 @@ var require_x509 = __commonJS({ pki2.certificationRequestFromPem = function(pem, computeHash, strict) { var msg = forge.pem.decode(pem)[0]; if (msg.type !== "CERTIFICATE REQUEST") { - var error4 = new Error('Could not convert certification request from PEM; PEM header type is not "CERTIFICATE REQUEST".'); - error4.headerType = msg.type; - throw error4; + var error3 = new Error('Could not convert certification request from PEM; PEM header type is not "CERTIFICATE REQUEST".'); + error3.headerType = msg.type; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { throw new Error("Could not convert certification request from PEM; PEM is encrypted."); @@ -33934,9 +33934,9 @@ var require_x509 = __commonJS({ cert.md = md || forge.md.sha1.create(); var algorithmOid = oids[cert.md.algorithm + "WithRSAEncryption"]; if (!algorithmOid) { - var error4 = new Error("Could not compute certificate digest. Unknown message digest algorithm OID."); - error4.algorithm = cert.md.algorithm; - throw error4; + var error3 = new Error("Could not compute certificate digest. Unknown message digest algorithm OID."); + error3.algorithm = cert.md.algorithm; + throw error3; } cert.signatureOid = cert.siginfo.algorithmOid = algorithmOid; cert.tbsCertificate = pki2.getTBSCertificate(cert); @@ -33949,12 +33949,12 @@ var require_x509 = __commonJS({ if (!cert.issued(child)) { var issuer = child.issuer; var subject = cert.subject; - var error4 = new Error( + var error3 = new Error( "The parent certificate did not issue the given child certificate; the child certificate's issuer does not match the parent's subject." ); - error4.expectedIssuer = subject.attributes; - error4.actualIssuer = issuer.attributes; - throw error4; + error3.expectedIssuer = subject.attributes; + error3.actualIssuer = issuer.attributes; + throw error3; } var md = child.md; if (md === null) { @@ -34017,9 +34017,9 @@ var require_x509 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, x509CertificateValidator, capture, errors)) { - var error4 = new Error("Cannot read X.509 certificate. ASN.1 object is not an X509v3 Certificate."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read X.509 certificate. ASN.1 object is not an X509v3 Certificate."); + error3.errors = errors; + throw error3; } var oid = asn1.derToOid(capture.publicKeyOid); if (oid !== pki2.oids.rsaEncryption) { @@ -34234,9 +34234,9 @@ var require_x509 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, certificationRequestValidator, capture, errors)) { - var error4 = new Error("Cannot read PKCS#10 certificate request. ASN.1 object is not a PKCS#10 CertificationRequest."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read PKCS#10 certificate request. ASN.1 object is not a PKCS#10 CertificationRequest."); + error3.errors = errors; + throw error3; } var oid = asn1.derToOid(capture.publicKeyOid); if (oid !== pki2.oids.rsaEncryption) { @@ -34332,9 +34332,9 @@ var require_x509 = __commonJS({ csr.md = md || forge.md.sha1.create(); var algorithmOid = oids[csr.md.algorithm + "WithRSAEncryption"]; if (!algorithmOid) { - var error4 = new Error("Could not compute certification request digest. Unknown message digest algorithm OID."); - error4.algorithm = csr.md.algorithm; - throw error4; + var error3 = new Error("Could not compute certification request digest. Unknown message digest algorithm OID."); + error3.algorithm = csr.md.algorithm; + throw error3; } csr.signatureOid = csr.siginfo.algorithmOid = algorithmOid; csr.certificationRequestInfo = pki2.getCertificationRequestInfo(csr); @@ -34416,9 +34416,9 @@ var require_x509 = __commonJS({ if (attr.name && attr.name in pki2.oids) { attr.type = pki2.oids[attr.name]; } else { - var error4 = new Error("Attribute type not specified."); - error4.attribute = attr; - throw error4; + var error3 = new Error("Attribute type not specified."); + error3.attribute = attr; + throw error3; } } if (typeof attr.shortName === "undefined") { @@ -34439,9 +34439,9 @@ var require_x509 = __commonJS({ } } if (typeof attr.value === "undefined") { - var error4 = new Error("Attribute value not specified."); - error4.attribute = attr; - throw error4; + var error3 = new Error("Attribute value not specified."); + error3.attribute = attr; + throw error3; } } } @@ -34456,9 +34456,9 @@ var require_x509 = __commonJS({ if (e.name && e.name in pki2.oids) { e.id = pki2.oids[e.name]; } else { - var error4 = new Error("Extension ID not specified."); - error4.extension = e; - throw error4; + var error3 = new Error("Extension ID not specified."); + error3.extension = e; + throw error3; } } if (typeof e.value !== "undefined") { @@ -34621,11 +34621,11 @@ var require_x509 = __commonJS({ if (altName.type === 7 && altName.ip) { value = forge.util.bytesFromIP(altName.ip); if (value === null) { - var error4 = new Error( + var error3 = new Error( 'Extension "ip" value is not a valid IPv4 or IPv6 address.' ); - error4.extension = e; - throw error4; + error3.extension = e; + throw error3; } } else if (altName.type === 8) { if (altName.oid) { @@ -34707,11 +34707,11 @@ var require_x509 = __commonJS({ if (altName.type === 7 && altName.ip) { value = forge.util.bytesFromIP(altName.ip); if (value === null) { - var error4 = new Error( + var error3 = new Error( 'Extension "ip" value is not a valid IPv4 or IPv6 address.' ); - error4.extension = e; - throw error4; + error3.extension = e; + throw error3; } } else if (altName.type === 8) { if (altName.oid) { @@ -34736,9 +34736,9 @@ var require_x509 = __commonJS({ seq2.push(subSeq); } if (typeof e.value === "undefined") { - var error4 = new Error("Extension value not specified."); - error4.extension = e; - throw error4; + var error3 = new Error("Extension value not specified."); + error3.extension = e; + throw error3; } return e; } @@ -35175,7 +35175,7 @@ var require_x509 = __commonJS({ validityCheckDate = /* @__PURE__ */ new Date(); } var first = true; - var error4 = null; + var error3 = null; var depth = 0; do { var cert = chain.shift(); @@ -35183,7 +35183,7 @@ var require_x509 = __commonJS({ var selfSigned = false; if (validityCheckDate) { if (validityCheckDate < cert.validity.notBefore || validityCheckDate > cert.validity.notAfter) { - error4 = { + error3 = { message: "Certificate is not valid yet or has expired.", error: pki2.certificateError.certificate_expired, notBefore: cert.validity.notBefore, @@ -35194,7 +35194,7 @@ var require_x509 = __commonJS({ }; } } - if (error4 === null) { + if (error3 === null) { parent = chain[0] || caStore.getIssuer(cert); if (parent === null) { if (cert.isIssuer(cert)) { @@ -35216,74 +35216,74 @@ var require_x509 = __commonJS({ } } if (!verified) { - error4 = { + error3 = { message: "Certificate signature is invalid.", error: pki2.certificateError.bad_certificate }; } } - if (error4 === null && (!parent || selfSigned) && !caStore.hasCertificate(cert)) { - error4 = { + if (error3 === null && (!parent || selfSigned) && !caStore.hasCertificate(cert)) { + error3 = { message: "Certificate is not trusted.", error: pki2.certificateError.unknown_ca }; } } - if (error4 === null && parent && !cert.isIssuer(parent)) { - error4 = { + if (error3 === null && parent && !cert.isIssuer(parent)) { + error3 = { message: "Certificate issuer is invalid.", error: pki2.certificateError.bad_certificate }; } - if (error4 === null) { + if (error3 === null) { var se = { keyUsage: true, basicConstraints: true }; - for (var i = 0; error4 === null && i < cert.extensions.length; ++i) { + for (var i = 0; error3 === null && i < cert.extensions.length; ++i) { var ext = cert.extensions[i]; if (ext.critical && !(ext.name in se)) { - error4 = { + error3 = { message: "Certificate has an unsupported critical extension.", error: pki2.certificateError.unsupported_certificate }; } } } - if (error4 === null && (!first || chain.length === 0 && (!parent || selfSigned))) { + if (error3 === null && (!first || chain.length === 0 && (!parent || selfSigned))) { var bcExt = cert.getExtension("basicConstraints"); var keyUsageExt = cert.getExtension("keyUsage"); if (keyUsageExt !== null) { if (!keyUsageExt.keyCertSign || bcExt === null) { - error4 = { + error3 = { message: "Certificate keyUsage or basicConstraints conflict or indicate that the certificate is not a CA. If the certificate is the only one in the chain or isn't the first then the certificate must be a valid CA.", error: pki2.certificateError.bad_certificate }; } } - if (error4 === null && bcExt !== null && !bcExt.cA) { - error4 = { + if (error3 === null && bcExt !== null && !bcExt.cA) { + error3 = { message: "Certificate basicConstraints indicates the certificate is not a CA.", error: pki2.certificateError.bad_certificate }; } - if (error4 === null && keyUsageExt !== null && "pathLenConstraint" in bcExt) { + if (error3 === null && keyUsageExt !== null && "pathLenConstraint" in bcExt) { var pathLen = depth - 1; if (pathLen > bcExt.pathLenConstraint) { - error4 = { + error3 = { message: "Certificate basicConstraints pathLenConstraint violated.", error: pki2.certificateError.bad_certificate }; } } } - var vfd = error4 === null ? true : error4.error; + var vfd = error3 === null ? true : error3.error; var ret = options.verify ? options.verify(vfd, depth, certs) : vfd; if (ret === true) { - error4 = null; + error3 = null; } else { if (vfd === true) { - error4 = { + error3 = { message: "The application rejected the certificate.", error: pki2.certificateError.bad_certificate }; @@ -35291,16 +35291,16 @@ var require_x509 = __commonJS({ if (ret || ret === 0) { if (typeof ret === "object" && !forge.util.isArray(ret)) { if (ret.message) { - error4.message = ret.message; + error3.message = ret.message; } if (ret.error) { - error4.error = ret.error; + error3.error = ret.error; } } else if (typeof ret === "string") { - error4.error = ret; + error3.error = ret; } } - throw error4; + throw error3; } first = false; ++depth; @@ -35513,9 +35513,9 @@ var require_pkcs12 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, pfxValidator, capture, errors)) { - var error4 = new Error("Cannot read PKCS#12 PFX. ASN.1 object is not an PKCS#12 PFX."); - error4.errors = error4; - throw error4; + var error3 = new Error("Cannot read PKCS#12 PFX. ASN.1 object is not an PKCS#12 PFX."); + error3.errors = error3; + throw error3; } var pfx = { version: capture.version.charCodeAt(0), @@ -35605,14 +35605,14 @@ var require_pkcs12 = __commonJS({ } }; if (capture.version.charCodeAt(0) !== 3) { - var error4 = new Error("PKCS#12 PFX of version other than 3 not supported."); - error4.version = capture.version.charCodeAt(0); - throw error4; + var error3 = new Error("PKCS#12 PFX of version other than 3 not supported."); + error3.version = capture.version.charCodeAt(0); + throw error3; } if (asn1.derToOid(capture.contentType) !== pki2.oids.data) { - var error4 = new Error("Only PKCS#12 PFX in password integrity mode supported."); - error4.oid = asn1.derToOid(capture.contentType); - throw error4; + var error3 = new Error("Only PKCS#12 PFX in password integrity mode supported."); + error3.oid = asn1.derToOid(capture.contentType); + throw error3; } var data = capture.content.value[0]; if (data.tagClass !== asn1.Class.UNIVERSAL || data.type !== asn1.Type.OCTETSTRING) { @@ -35690,9 +35690,9 @@ var require_pkcs12 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(contentInfo, contentInfoValidator, capture, errors)) { - var error4 = new Error("Cannot read ContentInfo."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read ContentInfo."); + error3.errors = errors; + throw error3; } var obj = { encrypted: false @@ -35711,9 +35711,9 @@ var require_pkcs12 = __commonJS({ obj.encrypted = true; break; default: - var error4 = new Error("Unsupported PKCS#12 contentType."); - error4.contentType = asn1.derToOid(capture.contentType); - throw error4; + var error3 = new Error("Unsupported PKCS#12 contentType."); + error3.contentType = asn1.derToOid(capture.contentType); + throw error3; } obj.safeBags = _decodeSafeContents(safeContents, strict, password); pfx.safeContents.push(obj); @@ -35728,17 +35728,17 @@ var require_pkcs12 = __commonJS({ capture, errors )) { - var error4 = new Error("Cannot read EncryptedContentInfo."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read EncryptedContentInfo."); + error3.errors = errors; + throw error3; } var oid = asn1.derToOid(capture.contentType); if (oid !== pki2.oids.data) { - var error4 = new Error( + var error3 = new Error( "PKCS#12 EncryptedContentInfo ContentType is not Data." ); - error4.oid = oid; - throw error4; + error3.oid = oid; + throw error3; } oid = asn1.derToOid(capture.encAlgorithm); var cipher = pki2.pbe.getCipher(oid, capture.encParameter, password); @@ -35766,9 +35766,9 @@ var require_pkcs12 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(safeBag, safeBagValidator, capture, errors)) { - var error4 = new Error("Cannot read SafeBag."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read SafeBag."); + error3.errors = errors; + throw error3; } var bag = { type: asn1.derToOid(capture.bagId), @@ -35799,11 +35799,11 @@ var require_pkcs12 = __commonJS({ validator = certBagValidator; decoder = function() { if (asn1.derToOid(capture.certId) !== pki2.oids.x509Certificate) { - var error5 = new Error( + var error4 = new Error( "Unsupported certificate type, only X.509 supported." ); - error5.oid = asn1.derToOid(capture.certId); - throw error5; + error4.oid = asn1.derToOid(capture.certId); + throw error4; } var certAsn1 = asn1.fromDer(capture.cert, strict); try { @@ -35815,14 +35815,14 @@ var require_pkcs12 = __commonJS({ }; break; default: - var error4 = new Error("Unsupported PKCS#12 SafeBag type."); - error4.oid = bag.type; - throw error4; + var error3 = new Error("Unsupported PKCS#12 SafeBag type."); + error3.oid = bag.type; + throw error3; } if (validator !== void 0 && !asn1.validate(bagAsn1, validator, capture, errors)) { - var error4 = new Error("Cannot read PKCS#12 " + validator.name); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read PKCS#12 " + validator.name); + error3.errors = errors; + throw error3; } decoder(); } @@ -35835,9 +35835,9 @@ var require_pkcs12 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(attributes[i], attributeValidator, capture, errors)) { - var error4 = new Error("Cannot read PKCS#12 BagAttribute."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read PKCS#12 BagAttribute."); + error3.errors = errors; + throw error3; } var oid = asn1.derToOid(capture.oid); if (pki2.oids[oid] === void 0) { @@ -36196,9 +36196,9 @@ var require_pki = __commonJS({ pki2.privateKeyFromPem = function(pem) { var msg = forge.pem.decode(pem)[0]; if (msg.type !== "PRIVATE KEY" && msg.type !== "RSA PRIVATE KEY") { - var error4 = new Error('Could not convert private key from PEM; PEM header type is not "PRIVATE KEY" or "RSA PRIVATE KEY".'); - error4.headerType = msg.type; - throw error4; + var error3 = new Error('Could not convert private key from PEM; PEM header type is not "PRIVATE KEY" or "RSA PRIVATE KEY".'); + error3.headerType = msg.type; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { throw new Error("Could not convert private key from PEM; PEM is encrypted."); @@ -36910,7 +36910,7 @@ var require_tls = __commonJS({ }); } if (c.serverCertificate === null) { - var error4 = { + var error3 = { message: "No server certificate provided. Not enough security.", send: true, alert: { @@ -36919,21 +36919,21 @@ var require_tls = __commonJS({ } }; var depth = 0; - var ret = c.verify(c, error4.alert.description, depth, []); + var ret = c.verify(c, error3.alert.description, depth, []); if (ret !== true) { if (ret || ret === 0) { if (typeof ret === "object" && !forge.util.isArray(ret)) { if (ret.message) { - error4.message = ret.message; + error3.message = ret.message; } if (ret.alert) { - error4.alert.description = ret.alert; + error3.alert.description = ret.alert; } } else if (typeof ret === "number") { - error4.alert.description = ret; + error3.alert.description = ret; } } - return c.error(c, error4); + return c.error(c, error3); } } if (c.session.certificateRequest !== null) { @@ -37581,9 +37581,9 @@ var require_tls = __commonJS({ for (var i = 0; i < cert.length; ++i) { var msg = forge.pem.decode(cert[i])[0]; if (msg.type !== "CERTIFICATE" && msg.type !== "X509 CERTIFICATE" && msg.type !== "TRUSTED CERTIFICATE") { - var error4 = new Error('Could not convert certificate from PEM; PEM header type is not "CERTIFICATE", "X509 CERTIFICATE", or "TRUSTED CERTIFICATE".'); - error4.headerType = msg.type; - throw error4; + var error3 = new Error('Could not convert certificate from PEM; PEM header type is not "CERTIFICATE", "X509 CERTIFICATE", or "TRUSTED CERTIFICATE".'); + error3.headerType = msg.type; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { throw new Error("Could not convert certificate from PEM; PEM is encrypted."); @@ -37809,8 +37809,8 @@ var require_tls = __commonJS({ c.records = []; return c.tlsDataReady(c); }; - var _certErrorToAlertDesc = function(error4) { - switch (error4) { + var _certErrorToAlertDesc = function(error3) { + switch (error3) { case true: return true; case forge.pki.certificateError.bad_certificate: @@ -37860,19 +37860,19 @@ var require_tls = __commonJS({ var ret = c.verify(c, vfd, depth, chain2); if (ret !== true) { if (typeof ret === "object" && !forge.util.isArray(ret)) { - var error4 = new Error("The application rejected the certificate."); - error4.send = true; - error4.alert = { + var error3 = new Error("The application rejected the certificate."); + error3.send = true; + error3.alert = { level: tls.Alert.Level.fatal, description: tls.Alert.Description.bad_certificate }; if (ret.message) { - error4.message = ret.message; + error3.message = ret.message; } if (ret.alert) { - error4.alert.description = ret.alert; + error3.alert.description = ret.alert; } - throw error4; + throw error3; } if (ret !== vfd) { ret = _alertDescToCertError(ret); @@ -38956,9 +38956,9 @@ var require_ed25519 = __commonJS({ var errors = []; var valid2 = forge.asn1.validate(obj, privateKeyValidator, capture, errors); if (!valid2) { - var error4 = new Error("Invalid Key."); - error4.errors = errors; - throw error4; + var error3 = new Error("Invalid Key."); + error3.errors = errors; + throw error3; } var oid = forge.asn1.derToOid(capture.privateKeyOid); var ed25519Oid = forge.oids.EdDSA25519; @@ -38977,9 +38977,9 @@ var require_ed25519 = __commonJS({ var errors = []; var valid2 = forge.asn1.validate(obj, publicKeyValidator, capture, errors); if (!valid2) { - var error4 = new Error("Invalid Key."); - error4.errors = errors; - throw error4; + var error3 = new Error("Invalid Key."); + error3.errors = errors; + throw error3; } var oid = forge.asn1.derToOid(capture.publicKeyOid); var ed25519Oid = forge.oids.EdDSA25519; @@ -40265,9 +40265,9 @@ var require_pkcs7 = __commonJS({ p7.messageFromPem = function(pem) { var msg = forge.pem.decode(pem)[0]; if (msg.type !== "PKCS7") { - var error4 = new Error('Could not convert PKCS#7 message from PEM; PEM header type is not "PKCS#7".'); - error4.headerType = msg.type; - throw error4; + var error3 = new Error('Could not convert PKCS#7 message from PEM; PEM header type is not "PKCS#7".'); + error3.headerType = msg.type; + throw error3; } if (msg.procType && msg.procType.type === "ENCRYPTED") { throw new Error("Could not convert PKCS#7 message from PEM; PEM is encrypted."); @@ -40286,9 +40286,9 @@ var require_pkcs7 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, p7.asn1.contentInfoValidator, capture, errors)) { - var error4 = new Error("Cannot read PKCS#7 message. ASN.1 object is not an PKCS#7 ContentInfo."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read PKCS#7 message. ASN.1 object is not an PKCS#7 ContentInfo."); + error3.errors = errors; + throw error3; } var contentType = asn1.derToOid(capture.contentType); var msg; @@ -40919,9 +40919,9 @@ var require_pkcs7 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, p7.asn1.recipientInfoValidator, capture, errors)) { - var error4 = new Error("Cannot read PKCS#7 RecipientInfo. ASN.1 object is not an PKCS#7 RecipientInfo."); - error4.errors = errors; - throw error4; + var error3 = new Error("Cannot read PKCS#7 RecipientInfo. ASN.1 object is not an PKCS#7 RecipientInfo."); + error3.errors = errors; + throw error3; } return { version: capture.version.charCodeAt(0), @@ -41162,9 +41162,9 @@ var require_pkcs7 = __commonJS({ var capture = {}; var errors = []; if (!asn1.validate(obj, validator, capture, errors)) { - var error4 = new Error("Cannot read PKCS#7 message. ASN.1 object is not a supported PKCS#7 message."); - error4.errors = error4; - throw error4; + var error3 = new Error("Cannot read PKCS#7 message. ASN.1 object is not a supported PKCS#7 message."); + error3.errors = error3; + throw error3; } var contentType = asn1.derToOid(capture.contentType); if (contentType !== forge.pki.oids.data) { @@ -42333,8 +42333,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -43066,7 +43066,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -43075,7 +43075,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -43085,17 +43085,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -43714,7 +43714,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -43723,7 +43723,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -43733,17 +43733,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -46415,9 +46415,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -47285,7 +47285,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { @@ -47320,23 +47320,21 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -47344,10 +47342,10 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", @@ -47357,9 +47355,9 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -47383,7 +47381,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -47569,8 +47568,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -47580,8 +47579,8 @@ var require_light = __commonJS({ return (await Promise.all(promises)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -47693,10 +47692,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -47730,7 +47729,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -47748,24 +47747,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } } doExpire(clearGlobalState, run, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -47775,7 +47774,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -48054,7 +48053,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve2, returned, task; + var args, cb, error3, reject, resolve2, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve2, reject } = this._queue.shift()); @@ -48065,9 +48064,9 @@ var require_light = __commonJS({ return resolve2(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -48201,8 +48200,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -48535,14 +48534,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -48838,24 +48837,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -48871,11 +48870,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -48896,12 +48895,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -53593,7 +53592,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core12 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob = __importStar4(require_glob()); var io4 = __importStar4(require_io4()); var crypto = __importStar4(require("crypto")); @@ -53677,7 +53676,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core12.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -55743,8 +55742,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -55978,9 +55977,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -57311,14 +57310,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -57328,7 +57327,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -57336,8 +57335,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -57557,7 +57556,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -58767,14 +58766,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -59446,11 +59445,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -59465,8 +59464,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -59960,8 +59959,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -60195,9 +60194,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -60697,8 +60696,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -60932,9 +60931,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -61997,9 +61996,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -62047,13 +62046,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -62073,21 +62072,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -62252,8 +62251,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -62659,16 +62658,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -63212,10 +63211,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -65294,12 +65293,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -65564,16 +65563,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -65698,7 +65697,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -65864,7 +65863,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -66105,9 +66104,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -66815,7 +66814,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -67878,26 +67877,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -67938,12 +67937,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -67951,13 +67950,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -67966,7 +67965,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -82572,8 +82571,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -82607,10 +82606,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -84194,8 +84193,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -84210,9 +84209,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve2, reject) => { this.emitter.on("finish", resolve2); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -85313,8 +85312,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -85398,7 +85397,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -88550,7 +88549,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -89917,9 +89916,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core12.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core12.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -90033,12 +90032,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -90072,13 +90071,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -90894,8 +90893,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -92741,8 +92740,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -94313,12 +94312,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -94338,12 +94337,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -94807,8 +94806,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -94819,8 +94818,8 @@ var require_test_transport = __commonJS({ stream.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } } @@ -95883,8 +95882,8 @@ var require_util11 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -95980,8 +95979,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -96012,18 +96011,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -96291,8 +96290,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -96493,22 +96492,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core12.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core12.error(`Failed to restore: ${error4.message}`); + core12.error(`Failed to restore: ${error3.message}`); } else { - core12.warning(`Failed to restore: ${error4.message}`); + core12.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -96563,15 +96562,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core12.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core12.error(`Failed to restore: ${error4.message}`); + core12.error(`Failed to restore: ${error3.message}`); } else { - core12.warning(`Failed to restore: ${error4.message}`); + core12.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -96579,8 +96578,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -96642,10 +96641,10 @@ var require_cache3 = __commonJS({ } core12.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core12.info(`Failed to save: ${typedError.message}`); } else { @@ -96658,8 +96657,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -96704,8 +96703,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core12.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core12.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core12.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -96724,10 +96723,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core12.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -96742,8 +96741,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -96768,7 +96767,6 @@ var io2 = __toESM(require_io3()); // src/util.ts var fsPromises = __toESM(require("fs/promises")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io3()); // node_modules/get-folder-size/index.js @@ -96786,14 +96784,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs.lstat(itemPath, { bigint: true }) : await fs.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs.lstat(itemPath, { bigint: true }) : await fs.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs.readdir(itemPath) : await fs.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs.readdir(itemPath) : await fs.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -96804,13 +96802,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -97690,6 +97688,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -97809,7 +97819,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -97849,16 +97859,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -99466,11 +99467,11 @@ function getTestingEnvironment() { } return testingEnvironment; } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } async function checkDiskUsage(logger) { try { @@ -99493,9 +99494,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -99683,8 +99684,8 @@ function getActionsLogger() { var core7 = __toESM(require_core()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/languages.ts var KnownLanguage = /* @__PURE__ */ ((KnownLanguage2) => { @@ -99895,13 +99896,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core9.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -99966,7 +99967,7 @@ async function getRef() { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; @@ -100034,6 +100035,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -100145,21 +100151,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, @@ -100172,6 +100178,8 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, @@ -100204,9 +100212,9 @@ function isFirstPartyAnalysis(actionName) { } return process.env["CODEQL_ACTION_INIT_HAS_RUN" /* INIT_ACTION_HAS_RUN */] === "true"; } -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -100481,11 +100489,11 @@ async function runWrapper() { logger ); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); - core11.setFailed(`start-proxy action failed: ${error4.message}`); + const error3 = wrapError(unwrappedError); + core11.setFailed(`start-proxy action failed: ${error3.message}`); const errorStatusReportBase = await createStatusReportBase( "start-proxy" /* StartProxy */, - getActionsStatus(error4), + getActionsStatus(error3), startedAt, { languages: language && [language] @@ -100517,8 +100525,8 @@ async function startProxy(binPath, config, logFilePath, logger) { if (subprocess.pid) { core11.saveState("proxy-process-pid", `${subprocess.pid}`); } - subprocess.on("error", (error4) => { - subprocessError = error4; + subprocess.on("error", (error3) => { + subprocessError = error3; }); subprocess.on("exit", (code) => { if (code !== 0) { @@ -100585,5 +100593,5 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/upload-lib.js b/lib/upload-lib.js index a0d12250c..f98674827 100644 --- a/lib/upload-lib.js +++ b/lib/upload-lib.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug4("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve6(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug4; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning8(message, properties = {}) { + exports2.error = error3; + function warning9(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning8; + exports2.warning = warning9; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -22042,8 +22042,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -22775,7 +22775,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -22784,7 +22784,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22794,17 +22794,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -23423,7 +23423,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -23432,7 +23432,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -23442,17 +23442,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -26124,9 +26124,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url2 = ""; return { value: { @@ -28924,7 +28924,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { @@ -28959,23 +28959,21 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -28983,10 +28981,10 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", @@ -28996,9 +28994,9 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -29022,7 +29020,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -29208,8 +29207,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -29219,8 +29218,8 @@ var require_light = __commonJS({ return (await Promise.all(promises3)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -29332,10 +29331,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -29369,7 +29368,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -29387,24 +29386,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } } doExpire(clearGlobalState, run, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -29414,7 +29413,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -29693,7 +29692,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve6, returned, task; + var args, cb, error3, reject, resolve6, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve6, reject } = this._queue.shift()); @@ -29704,9 +29703,9 @@ var require_light = __commonJS({ return resolve6(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -29840,8 +29839,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -30174,14 +30173,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -30477,24 +30476,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -30510,11 +30509,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -30535,12 +30534,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -33935,7 +33934,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core12 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob = __importStar4(require_glob()); var io6 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -34019,7 +34018,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core12.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -36085,8 +36084,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -36320,9 +36319,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -37653,14 +37652,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -37670,7 +37669,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -37678,8 +37677,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -37899,7 +37898,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -39109,14 +39108,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -39788,11 +39787,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -39807,8 +39806,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -40302,8 +40301,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -40537,9 +40536,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -41039,8 +41038,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -41274,9 +41273,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -42339,9 +42338,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -42389,13 +42388,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -42415,21 +42414,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -42594,8 +42593,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -43001,16 +43000,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -43554,10 +43553,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -45636,12 +45635,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -45906,16 +45905,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -46040,7 +46039,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -46206,7 +46205,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -46447,9 +46446,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -47157,7 +47156,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -48220,26 +48219,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -48280,12 +48279,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -48293,13 +48292,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -48308,7 +48307,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -62914,8 +62913,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -62949,10 +62948,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -64536,8 +64535,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -64552,9 +64551,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve6, reject) => { this.emitter.on("finish", resolve6); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -65655,8 +65654,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -65740,7 +65739,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -68892,7 +68891,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -70259,9 +70258,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core12.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core12.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -70375,12 +70374,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -70414,13 +70413,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -71236,8 +71235,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -73083,8 +73082,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -74655,12 +74654,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -74680,12 +74679,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -75149,8 +75148,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -75161,8 +75160,8 @@ var require_test_transport = __commonJS({ stream2.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } } @@ -76225,8 +76224,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url2} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url2} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -76322,8 +76321,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url2, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -76354,18 +76353,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -76633,8 +76632,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -76835,22 +76834,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core12.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core12.error(`Failed to restore: ${error4.message}`); + core12.error(`Failed to restore: ${error3.message}`); } else { - core12.warning(`Failed to restore: ${error4.message}`); + core12.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76905,15 +76904,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core12.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core12.error(`Failed to restore: ${error4.message}`); + core12.error(`Failed to restore: ${error3.message}`); } else { - core12.warning(`Failed to restore: ${error4.message}`); + core12.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -76921,8 +76920,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -76984,10 +76983,10 @@ var require_cache3 = __commonJS({ } core12.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core12.info(`Failed to save: ${typedError.message}`); } else { @@ -77000,8 +76999,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -77046,8 +77045,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core12.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core12.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core12.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -77066,10 +77065,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core12.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -77084,8 +77083,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core12.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core12.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79810,7 @@ var require_debug2 = __commonJS({ if (!debug4) { try { debug4 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug4 !== "function") { debug4 = function() { @@ -79844,8 +79843,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79918,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80087,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80289,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -83222,7 +83221,6 @@ __export(upload_lib_exports, { buildPayload: () => buildPayload, findSarifFilesInDir: () => findSarifFilesInDir, getGroupedSarifFilePaths: () => getGroupedSarifFilePaths, - getSarifFilePaths: () => getSarifFilePaths, populateRunAutomationDetails: () => populateRunAutomationDetails, postProcessSarifFiles: () => postProcessSarifFiles, readSarifFile: () => readSarifFile, @@ -83258,7 +83256,6 @@ var io2 = __toESM(require_io2()); var fs = __toESM(require("fs")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js @@ -83276,14 +83273,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs12.lstat(itemPath, { bigint: true }) : await fs12.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs12.lstat(itemPath, { bigint: true }) : await fs12.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs12.readdir(itemPath) : await fs12.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs12.readdir(itemPath) : await fs12.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -83294,13 +83291,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -84180,6 +84177,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -84299,7 +84308,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -84339,16 +84348,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -85913,9 +85913,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -86051,11 +86051,11 @@ function parseMatrixInput(matrixInput) { } return JSON.parse(matrixInput); } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } function satisfiesGHESVersion(ghesVersion, range, defaultIfInvalid) { const semverVersion = semver.coerce(ghesVersion); @@ -86489,19 +86489,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -86517,9 +86517,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -86724,8 +86724,8 @@ var path4 = __toESM(require("path")); var semver4 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); @@ -86754,13 +86754,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -86944,7 +86944,7 @@ function formatDuration(durationMs) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { @@ -87072,6 +87072,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -87183,21 +87188,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, @@ -87228,6 +87233,8 @@ ${jsonContents}` var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, @@ -90308,15 +90315,15 @@ function validateSarifFileSchema(sarif, sarifFilePath, logger) { const warnings = (result.errors ?? []).filter( (err) => err.name === "format" && typeof err.argument === "string" && warningAttributes.includes(err.argument) ); - for (const warning8 of warnings) { + for (const warning9 of warnings) { logger.info( - `Warning: '${warning8.instance}' is not a valid URI in '${warning8.property}'.` + `Warning: '${warning9.instance}' is not a valid URI in '${warning9.property}'.` ); } if (errors.length > 0) { - for (const error4 of errors) { - logger.startGroup(`Error details: ${error4.stack}`); - logger.info(JSON.stringify(error4, null, 2)); + for (const error3 of errors) { + logger.startGroup(`Error details: ${error3.stack}`); + logger.info(JSON.stringify(error3, null, 2)); logger.endGroup(); } const sarifErrors = errors.map((e) => `- ${e.stack}`); @@ -90569,10 +90576,10 @@ function shouldConsiderConfigurationError(processingErrors) { } function shouldConsiderInvalidRequest(processingErrors) { return processingErrors.every( - (error4) => error4.startsWith("rejecting SARIF") || error4.startsWith("an invalid URI was provided as a SARIF location") || error4.startsWith("locationFromSarifResult: expected artifact location") || error4.startsWith( + (error3) => error3.startsWith("rejecting SARIF") || error3.startsWith("an invalid URI was provided as a SARIF location") || error3.startsWith("locationFromSarifResult: expected artifact location") || error3.startsWith( "could not convert rules: invalid security severity value, is not a number" ) || /^SARIF URI scheme [^\s]* did not match the checkout URI scheme [^\s]*/.test( - error4 + error3 ) ); } @@ -90655,7 +90662,6 @@ function filterAlertsByDiffRange(logger, sarif) { buildPayload, findSarifFilesInDir, getGroupedSarifFilePaths, - getSarifFilePaths, populateRunAutomationDetails, postProcessSarifFiles, readSarifFile, @@ -90680,7 +90686,7 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) long/index.js: (** diff --git a/lib/upload-sarif-action-post.js b/lib/upload-sarif-action-post.js index 1d6354708..6256ed8ed 100644 --- a/lib/upload-sarif-action-post.js +++ b/lib/upload-sarif-action-post.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug4("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay === "number" && delay > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve2(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug4; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning9(message, properties = {}) { + exports2.error = error3; + function warning10(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning9; + exports2.warning = warning10; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url = ""; return { value: { @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { @@ -27662,23 +27662,21 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -27686,10 +27684,10 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", @@ -27699,9 +27697,9 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27723,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -27911,8 +27910,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27921,8 @@ var require_light = __commonJS({ return (await Promise.all(promises2)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28034,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28071,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28089,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } } doExpire(clearGlobalState, run, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28116,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28395,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve2, returned, task; + var args, cb, error3, reject, resolve2, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve2, reject } = this._queue.shift()); @@ -28407,9 +28406,9 @@ var require_light = __commonJS({ return resolve2(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28542,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28876,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29179,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29212,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29237,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -31147,8 +31146,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -33460,12 +33459,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -33485,12 +33484,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -33954,8 +33953,8 @@ var require_test_transport = __commonJS({ } try { yield delay(this.responseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -33966,8 +33965,8 @@ var require_test_transport = __commonJS({ stream.notifyMessage(msg); try { yield delay(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream.notifyError(error4); + } catch (error3) { + stream.notifyError(error3); return; } } @@ -36797,8 +36796,8 @@ var require_util8 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } function maskSecretUrls(body) { @@ -36891,8 +36890,8 @@ var require_artifact_twirp_client2 = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -36923,18 +36922,18 @@ var require_artifact_twirp_client2 = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -38273,8 +38272,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -38508,9 +38507,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -39841,14 +39840,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -39858,7 +39857,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -39866,8 +39865,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -40087,7 +40086,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -41297,14 +41296,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -41976,11 +41975,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -41995,8 +41994,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -42490,8 +42489,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -42725,9 +42724,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -43227,8 +43226,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -43462,9 +43461,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -44527,9 +44526,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -44577,13 +44576,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -44603,21 +44602,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -44782,8 +44781,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -45189,16 +45188,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -45742,10 +45741,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -47824,12 +47823,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -48094,16 +48093,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -48228,7 +48227,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -48394,7 +48393,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -48635,9 +48634,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -49345,7 +49344,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -50408,26 +50407,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -50468,12 +50467,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -50481,13 +50480,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -50496,7 +50495,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -65102,8 +65101,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -65137,10 +65136,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -66724,8 +66723,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -66740,9 +66739,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve2, reject) => { this.emitter.on("finish", resolve2); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -67843,8 +67842,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -67928,7 +67927,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -71080,7 +71079,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -72327,11 +72326,11 @@ var require_blob_upload = __commonJS({ blockBlobClient.uploadStream(uploadStream, bufferSize, maxConcurrency, options), chunkTimer((0, config_1.getUploadChunkTimeout)()) ]); - } catch (error4) { - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + } catch (error3) { + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } - throw error4; + throw error3; } finally { abortController.abort(); } @@ -73407,9 +73406,9 @@ var require_async = __commonJS({ invokeCallback(callback, err && (err instanceof Error || err.message) ? err : new Error(err)); }); } - function invokeCallback(callback, error4, value) { + function invokeCallback(callback, error3, value) { try { - callback(error4, value); + callback(error3, value); } catch (err) { setImmediate$1((e) => { throw e; @@ -74715,10 +74714,10 @@ var require_async = __commonJS({ function reflect(fn) { var _fn = wrapAsync(fn); return initialParams(function reflectOn(args, reflectCallback) { - args.push((error4, ...cbArgs) => { + args.push((error3, ...cbArgs) => { let retVal = {}; - if (error4) { - retVal.error = error4; + if (error3) { + retVal.error = error3; } if (cbArgs.length > 0) { var value = cbArgs; @@ -74874,13 +74873,13 @@ var require_async = __commonJS({ var timer; function timeoutCallback() { var name = asyncFn.name || "anonymous"; - var error4 = new Error('Callback function "' + name + '" timed out.'); - error4.code = "ETIMEDOUT"; + var error3 = new Error('Callback function "' + name + '" timed out.'); + error3.code = "ETIMEDOUT"; if (info7) { - error4.info = info7; + error3.info = info7; } timedOut = true; - callback(error4); + callback(error3); } args.push((...cbArgs) => { if (!timedOut) { @@ -74923,7 +74922,7 @@ var require_async = __commonJS({ return callback[PROMISE_SYMBOL]; } function tryEach(tasks, callback) { - var error4 = null; + var error3 = null; var result; return eachSeries$1(tasks, (task, taskCb) => { wrapAsync(task)((err, ...args) => { @@ -74933,10 +74932,10 @@ var require_async = __commonJS({ } else { result = args; } - error4 = err; + error3 = err; taskCb(err ? null : {}); }); - }, () => callback(error4, result)); + }, () => callback(error3, result)); } var tryEach$1 = awaitify(tryEach); function unmemoize(fn) { @@ -81416,19 +81415,19 @@ var require_from = __commonJS({ next(); } }; - readable._destroy = function(error4, cb) { + readable._destroy = function(error3, cb) { PromisePrototypeThen( - close(error4), - () => process2.nextTick(cb, error4), + close(error3), + () => process2.nextTick(cb, error3), // nextTick is here in case cb throws - (e) => process2.nextTick(cb, e || error4) + (e) => process2.nextTick(cb, e || error3) ); }; - async function close(error4) { - const hadError = error4 !== void 0 && error4 !== null; + async function close(error3) { + const hadError = error3 !== void 0 && error3 !== null; const hasThrow = typeof iterator.throw === "function"; if (hadError && hasThrow) { - const { value, done } = await iterator.throw(error4); + const { value, done } = await iterator.throw(error3); await value; if (done) { return; @@ -81636,12 +81635,12 @@ var require_readable3 = __commonJS({ this.destroy(err); }; Readable.prototype[SymbolAsyncDispose] = function() { - let error4; + let error3; if (!this.destroyed) { - error4 = this.readableEnded ? null : new AbortError(); - this.destroy(error4); + error3 = this.readableEnded ? null : new AbortError(); + this.destroy(error3); } - return new Promise2((resolve2, reject) => eos(this, (err) => err && err !== error4 ? reject(err) : resolve2(null))); + return new Promise2((resolve2, reject) => eos(this, (err) => err && err !== error3 ? reject(err) : resolve2(null))); }; Readable.prototype.push = function(chunk, encoding) { return readableAddChunk(this, chunk, encoding, false); @@ -82194,14 +82193,14 @@ var require_readable3 = __commonJS({ } } stream.on("readable", next); - let error4; + let error3; const cleanup = eos( stream, { writable: false }, (err) => { - error4 = err ? aggregateTwoErrors(error4, err) : null; + error3 = err ? aggregateTwoErrors(error3, err) : null; callback(); callback = nop; } @@ -82211,19 +82210,19 @@ var require_readable3 = __commonJS({ const chunk = stream.destroyed ? null : stream.read(); if (chunk !== null) { yield chunk; - } else if (error4) { - throw error4; - } else if (error4 === null) { + } else if (error3) { + throw error3; + } else if (error3 === null) { return; } else { await new Promise2(next); } } } catch (err) { - error4 = aggregateTwoErrors(error4, err); - throw error4; + error3 = aggregateTwoErrors(error3, err); + throw error3; } finally { - if ((error4 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error4 === void 0 || stream._readableState.autoDestroy)) { + if ((error3 || (options === null || options === void 0 ? void 0 : options.destroyOnReturn) !== false) && (error3 === void 0 || stream._readableState.autoDestroy)) { destroyImpl.destroyer(stream, null); } else { stream.off("readable", next); @@ -83716,11 +83715,11 @@ var require_pipeline3 = __commonJS({ yield* Readable.prototype[SymbolAsyncIterator].call(val2); } async function pumpToNode(iterable, writable, finish, { end }) { - let error4; + let error3; let onresolve = null; const resume = (err) => { if (err) { - error4 = err; + error3 = err; } if (onresolve) { const callback = onresolve; @@ -83729,12 +83728,12 @@ var require_pipeline3 = __commonJS({ } }; const wait = () => new Promise2((resolve2, reject) => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { onresolve = () => { - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve2(); } @@ -83764,7 +83763,7 @@ var require_pipeline3 = __commonJS({ } finish(); } catch (err) { - finish(error4 !== err ? aggregateTwoErrors(error4, err) : err); + finish(error3 !== err ? aggregateTwoErrors(error3, err) : err); } finally { cleanup(); writable.off("drain", resume); @@ -83818,7 +83817,7 @@ var require_pipeline3 = __commonJS({ if (outerSignal) { disposable = addAbortListener(outerSignal, abort); } - let error4; + let error3; let value; const destroys = []; let finishCount = 0; @@ -83827,23 +83826,23 @@ var require_pipeline3 = __commonJS({ } function finishImpl(err, final) { var _disposable; - if (err && (!error4 || error4.code === "ERR_STREAM_PREMATURE_CLOSE")) { - error4 = err; + if (err && (!error3 || error3.code === "ERR_STREAM_PREMATURE_CLOSE")) { + error3 = err; } - if (!error4 && !final) { + if (!error3 && !final) { return; } while (destroys.length) { - destroys.shift()(error4); + destroys.shift()(error3); } ; (_disposable = disposable) === null || _disposable === void 0 ? void 0 : _disposable[SymbolDispose](); ac.abort(); if (final) { - if (!error4) { + if (!error3) { lastStreamCleanup.forEach((fn) => fn()); } - process2.nextTick(callback, error4, value); + process2.nextTick(callback, error3, value); } } let ret; @@ -85693,52 +85692,128 @@ var require_isPlainObject = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/brace-expansion/index.js -var require_brace_expansion2 = __commonJS({ - "node_modules/archiver-utils/node_modules/brace-expansion/index.js"(exports2, module2) { - var balanced = require_balanced_match(); - module2.exports = expandTop; +// node_modules/@isaacs/balanced-match/dist/commonjs/index.js +var require_commonjs13 = __commonJS({ + "node_modules/@isaacs/balanced-match/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.range = exports2.balanced = void 0; + var balanced = (a, b, str2) => { + const ma = a instanceof RegExp ? maybeMatch(a, str2) : a; + const mb = b instanceof RegExp ? maybeMatch(b, str2) : b; + const r = ma !== null && mb != null && (0, exports2.range)(ma, mb, str2); + return r && { + start: r[0], + end: r[1], + pre: str2.slice(0, r[0]), + body: str2.slice(r[0] + ma.length, r[1]), + post: str2.slice(r[1] + mb.length) + }; + }; + exports2.balanced = balanced; + var maybeMatch = (reg, str2) => { + const m = str2.match(reg); + return m ? m[0] : null; + }; + var range = (a, b, str2) => { + let begs, beg, left, right = void 0, result; + let ai = str2.indexOf(a); + let bi = str2.indexOf(b, ai + 1); + let i = ai; + if (ai >= 0 && bi > 0) { + if (a === b) { + return [ai, bi]; + } + begs = []; + left = str2.length; + while (i >= 0 && !result) { + if (i === ai) { + begs.push(i); + ai = str2.indexOf(a, i + 1); + } else if (begs.length === 1) { + const r = begs.pop(); + if (r !== void 0) + result = [r, bi]; + } else { + beg = begs.pop(); + if (beg !== void 0 && beg < left) { + left = beg; + right = bi; + } + bi = str2.indexOf(b, i + 1); + } + i = ai < bi && ai >= 0 ? ai : bi; + } + if (begs.length && right !== void 0) { + result = [left, right]; + } + } + return result; + }; + exports2.range = range; + } +}); + +// node_modules/@isaacs/brace-expansion/dist/commonjs/index.js +var require_commonjs14 = __commonJS({ + "node_modules/@isaacs/brace-expansion/dist/commonjs/index.js"(exports2) { + "use strict"; + Object.defineProperty(exports2, "__esModule", { value: true }); + exports2.expand = expand; + var balanced_match_1 = require_commonjs13(); var escSlash = "\0SLASH" + Math.random() + "\0"; var escOpen = "\0OPEN" + Math.random() + "\0"; var escClose = "\0CLOSE" + Math.random() + "\0"; var escComma = "\0COMMA" + Math.random() + "\0"; var escPeriod = "\0PERIOD" + Math.random() + "\0"; + var escSlashPattern = new RegExp(escSlash, "g"); + var escOpenPattern = new RegExp(escOpen, "g"); + var escClosePattern = new RegExp(escClose, "g"); + var escCommaPattern = new RegExp(escComma, "g"); + var escPeriodPattern = new RegExp(escPeriod, "g"); + var slashPattern = /\\\\/g; + var openPattern = /\\{/g; + var closePattern = /\\}/g; + var commaPattern = /\\,/g; + var periodPattern = /\\./g; function numeric(str2) { - return parseInt(str2, 10) == str2 ? parseInt(str2, 10) : str2.charCodeAt(0); + return !isNaN(str2) ? parseInt(str2, 10) : str2.charCodeAt(0); } function escapeBraces(str2) { - return str2.split("\\\\").join(escSlash).split("\\{").join(escOpen).split("\\}").join(escClose).split("\\,").join(escComma).split("\\.").join(escPeriod); + return str2.replace(slashPattern, escSlash).replace(openPattern, escOpen).replace(closePattern, escClose).replace(commaPattern, escComma).replace(periodPattern, escPeriod); } function unescapeBraces(str2) { - return str2.split(escSlash).join("\\").split(escOpen).join("{").split(escClose).join("}").split(escComma).join(",").split(escPeriod).join("."); + return str2.replace(escSlashPattern, "\\").replace(escOpenPattern, "{").replace(escClosePattern, "}").replace(escCommaPattern, ",").replace(escPeriodPattern, "."); } function parseCommaParts(str2) { - if (!str2) + if (!str2) { return [""]; - var parts = []; - var m = balanced("{", "}", str2); - if (!m) + } + const parts = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) { return str2.split(","); - var pre = m.pre; - var body = m.body; - var post = m.post; - var p = pre.split(","); + } + const { pre, body, post } = m; + const p = pre.split(","); p[p.length - 1] += "{" + body + "}"; - var postParts = parseCommaParts(post); + const postParts = parseCommaParts(post); if (post.length) { + ; p[p.length - 1] += postParts.shift(); p.push.apply(p, postParts); } parts.push.apply(parts, p); return parts; } - function expandTop(str2) { - if (!str2) + function expand(str2) { + if (!str2) { return []; - if (str2.substr(0, 2) === "{}") { - str2 = "\\{\\}" + str2.substr(2); } - return expand(escapeBraces(str2), true).map(unescapeBraces); + if (str2.slice(0, 2) === "{}") { + str2 = "\\{\\}" + str2.slice(2); + } + return expand_(escapeBraces(str2), true).map(unescapeBraces); } function embrace(str2) { return "{" + str2 + "}"; @@ -85752,73 +85827,74 @@ var require_brace_expansion2 = __commonJS({ function gte5(i, y) { return i >= y; } - function expand(str2, isTop) { - var expansions = []; - var m = balanced("{", "}", str2); - if (!m) return [str2]; - var pre = m.pre; - var post = m.post.length ? expand(m.post, false) : [""]; + function expand_(str2, isTop) { + const expansions = []; + const m = (0, balanced_match_1.balanced)("{", "}", str2); + if (!m) + return [str2]; + const pre = m.pre; + const post = m.post.length ? expand_(m.post, false) : [""]; if (/\$$/.test(m.pre)) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + "{" + m.body + "}" + post[k]; + for (let k = 0; k < post.length; k++) { + const expansion = pre + "{" + m.body + "}" + post[k]; expansions.push(expansion); } } else { - var isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); - var isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); - var isSequence = isNumericSequence || isAlphaSequence; - var isOptions = m.body.indexOf(",") >= 0; + const isNumericSequence = /^-?\d+\.\.-?\d+(?:\.\.-?\d+)?$/.test(m.body); + const isAlphaSequence = /^[a-zA-Z]\.\.[a-zA-Z](?:\.\.-?\d+)?$/.test(m.body); + const isSequence = isNumericSequence || isAlphaSequence; + const isOptions = m.body.indexOf(",") >= 0; if (!isSequence && !isOptions) { if (m.post.match(/,(?!,).*\}/)) { str2 = m.pre + "{" + m.body + escClose + m.post; - return expand(str2); + return expand_(str2); } return [str2]; } - var n; + let n; if (isSequence) { n = m.body.split(/\.\./); } else { n = parseCommaParts(m.body); - if (n.length === 1) { - n = expand(n[0], false).map(embrace); + if (n.length === 1 && n[0] !== void 0) { + n = expand_(n[0], false).map(embrace); if (n.length === 1) { - return post.map(function(p) { - return m.pre + n[0] + p; - }); + return post.map((p) => m.pre + n[0] + p); } } } - var N; - if (isSequence) { - var x = numeric(n[0]); - var y = numeric(n[1]); - var width = Math.max(n[0].length, n[1].length); - var incr = n.length == 3 ? Math.abs(numeric(n[2])) : 1; - var test = lte; - var reverse = y < x; + let N; + if (isSequence && n[0] !== void 0 && n[1] !== void 0) { + const x = numeric(n[0]); + const y = numeric(n[1]); + const width = Math.max(n[0].length, n[1].length); + let incr = n.length === 3 && n[2] !== void 0 ? Math.abs(numeric(n[2])) : 1; + let test = lte; + const reverse = y < x; if (reverse) { incr *= -1; test = gte5; } - var pad = n.some(isPadded); + const pad = n.some(isPadded); N = []; - for (var i = x; test(i, y); i += incr) { - var c; + for (let i = x; test(i, y); i += incr) { + let c; if (isAlphaSequence) { c = String.fromCharCode(i); - if (c === "\\") + if (c === "\\") { c = ""; + } } else { c = String(i); if (pad) { - var need = width - c.length; + const need = width - c.length; if (need > 0) { - var z = new Array(need + 1).join("0"); - if (i < 0) + const z = new Array(need + 1).join("0"); + if (i < 0) { c = "-" + z + c.slice(1); - else + } else { c = z + c; + } } } } @@ -85826,15 +85902,16 @@ var require_brace_expansion2 = __commonJS({ } } else { N = []; - for (var j = 0; j < n.length; j++) { - N.push.apply(N, expand(n[j], false)); + for (let j = 0; j < n.length; j++) { + N.push.apply(N, expand_(n[j], false)); } } - for (var j = 0; j < N.length; j++) { - for (var k = 0; k < post.length; k++) { - var expansion = pre + N[j] + post[k]; - if (!isTop || isSequence || expansion) + for (let j = 0; j < N.length; j++) { + for (let k = 0; k < post.length; k++) { + const expansion = pre + N[j] + post[k]; + if (!isTop || isSequence || expansion) { expansions.push(expansion); + } } } } @@ -85843,9 +85920,9 @@ var require_brace_expansion2 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js var require_assert_valid_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/assert-valid-pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.assertValidPattern = void 0; @@ -85862,9 +85939,9 @@ var require_assert_valid_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js var require_brace_expressions = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/brace-expressions.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.parseClass = void 0; @@ -85979,22 +86056,25 @@ var require_brace_expressions = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js var require_unescape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/unescape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = void 0; - var unescape = (s, { windowsPathsNoEscape = false } = {}) => { - return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + var unescape = (s, { windowsPathsNoEscape = false, magicalBraces = true } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/\[([^\/\\])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\])\]/g, "$1$2").replace(/\\([^\/])/g, "$1"); + } + return windowsPathsNoEscape ? s.replace(/\[([^\/\\{}])\]/g, "$1") : s.replace(/((?!\\).|^)\[([^\/\\{}])\]/g, "$1$2").replace(/\\([^\/{}])/g, "$1"); }; exports2.unescape = unescape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js var require_ast = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/ast.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.AST = void 0; @@ -86350,7 +86430,7 @@ var require_ast = __commonJS({ if (this.#root === this) this.#fillNegs(); if (!this.type) { - const noEmpty = this.isStart() && this.isEnd(); + const noEmpty = this.isStart() && this.isEnd() && !this.#parts.some((s) => typeof s !== "string"); const src = this.#parts.map((p) => { const [re, _2, hasMagic, uflag] = typeof p === "string" ? _AST.#parseGlob(p, this.#hasMagic, noEmpty) : p.toRegExpSource(allowDot); this.#hasMagic = this.#hasMagic || hasMagic; @@ -86460,10 +86540,7 @@ var require_ast = __commonJS({ } } if (c === "*") { - if (noEmpty && glob2 === "*") - re += starNoEmpty; - else - re += star; + re += noEmpty && glob2 === "*" ? starNoEmpty : star; hasMagic = true; continue; } @@ -86481,29 +86558,29 @@ var require_ast = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js +// node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js var require_escape = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { + "node_modules/glob/node_modules/minimatch/dist/commonjs/escape.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.escape = void 0; - var escape = (s, { windowsPathsNoEscape = false } = {}) => { + var escape = (s, { windowsPathsNoEscape = false, magicalBraces = false } = {}) => { + if (magicalBraces) { + return windowsPathsNoEscape ? s.replace(/[?*()[\]{}]/g, "[$&]") : s.replace(/[?*()[\]\\{}]/g, "\\$&"); + } return windowsPathsNoEscape ? s.replace(/[?*()[\]]/g, "[$&]") : s.replace(/[?*()[\]\\]/g, "\\$&"); }; exports2.escape = escape; } }); -// node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js -var require_commonjs13 = __commonJS({ - "node_modules/archiver-utils/node_modules/minimatch/dist/commonjs/index.js"(exports2) { +// node_modules/glob/node_modules/minimatch/dist/commonjs/index.js +var require_commonjs15 = __commonJS({ + "node_modules/glob/node_modules/minimatch/dist/commonjs/index.js"(exports2) { "use strict"; - var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { - return mod && mod.__esModule ? mod : { "default": mod }; - }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.unescape = exports2.escape = exports2.AST = exports2.Minimatch = exports2.match = exports2.makeRe = exports2.braceExpand = exports2.defaults = exports2.filter = exports2.GLOBSTAR = exports2.sep = exports2.minimatch = void 0; - var brace_expansion_1 = __importDefault4(require_brace_expansion2()); + var brace_expansion_1 = require_commonjs14(); var assert_valid_pattern_js_1 = require_assert_valid_pattern(); var ast_js_1 = require_ast(); var escape_js_1 = require_escape(); @@ -86626,7 +86703,7 @@ var require_commonjs13 = __commonJS({ if (options.nobrace || !/\{(?:(?!\{).)*\}/.test(pattern)) { return [pattern]; } - return (0, brace_expansion_1.default)(pattern); + return (0, brace_expansion_1.expand)(pattern); }; exports2.braceExpand = braceExpand; exports2.minimatch.braceExpand = exports2.braceExpand; @@ -87150,16 +87227,27 @@ var require_commonjs13 = __commonJS({ pp[i] = twoStar; } } else if (next === void 0) { - pp[i - 1] = prev + "(?:\\/|" + twoStar + ")?"; + pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + ")?"; } else if (next !== exports2.GLOBSTAR) { pp[i - 1] = prev + "(?:\\/|\\/" + twoStar + "\\/)" + next; pp[i + 1] = exports2.GLOBSTAR; } }); - return pp.filter((p) => p !== exports2.GLOBSTAR).join("/"); + const filtered = pp.filter((p) => p !== exports2.GLOBSTAR); + if (this.partial && filtered.length >= 1) { + const prefixes = []; + for (let i = 1; i <= filtered.length; i++) { + prefixes.push(filtered.slice(0, i).join("/")); + } + return "(?:" + prefixes.join("|") + ")"; + } + return filtered.join("/"); }).join("|"); const [open, close] = set2.length > 1 ? ["(?:", ")"] : ["", ""]; re = "^" + open + re + close + "$"; + if (this.partial) { + re = "^(?:\\/|" + open + re.slice(1, -1) + close + ")$"; + } if (this.negate) re = "^(?!" + re + ").+$"; try { @@ -87246,9 +87334,9 @@ var require_commonjs13 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js -var require_commonjs14 = __commonJS({ - "node_modules/archiver-utils/node_modules/lru-cache/dist/commonjs/index.js"(exports2) { +// node_modules/lru-cache/dist/commonjs/index.js +var require_commonjs16 = __commonJS({ + "node_modules/lru-cache/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.LRUCache = void 0; @@ -87337,6 +87425,7 @@ var require_commonjs14 = __commonJS({ #max; #maxSize; #dispose; + #onInsert; #disposeAfter; #fetchMethod; #memoMethod; @@ -87418,6 +87507,7 @@ var require_commonjs14 = __commonJS({ #hasDispose; #hasFetchMethod; #hasDisposeAfter; + #hasOnInsert; /** * Do not call this method unless you need to inspect the * inner workings of the cache. If anything returned by this @@ -87494,6 +87584,12 @@ var require_commonjs14 = __commonJS({ get dispose() { return this.#dispose; } + /** + * {@link LRUCache.OptionsBase.onInsert} (read-only) + */ + get onInsert() { + return this.#onInsert; + } /** * {@link LRUCache.OptionsBase.disposeAfter} (read-only) */ @@ -87501,7 +87597,7 @@ var require_commonjs14 = __commonJS({ return this.#disposeAfter; } constructor(options) { - const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; + const { max = 0, ttl, ttlResolution = 1, ttlAutopurge, updateAgeOnGet, updateAgeOnHas, allowStale, dispose, onInsert, disposeAfter, noDisposeOnSet, noUpdateTTL, maxSize = 0, maxEntrySize = 0, sizeCalculation, fetchMethod, memoMethod, noDeleteOnFetchRejection, noDeleteOnStaleGet, allowStaleOnFetchRejection, allowStaleOnFetchAbort, ignoreFetchAbort } = options; if (max !== 0 && !isPosInt(max)) { throw new TypeError("max option must be a nonnegative integer"); } @@ -87543,6 +87639,9 @@ var require_commonjs14 = __commonJS({ if (typeof dispose === "function") { this.#dispose = dispose; } + if (typeof onInsert === "function") { + this.#onInsert = onInsert; + } if (typeof disposeAfter === "function") { this.#disposeAfter = disposeAfter; this.#disposed = []; @@ -87551,6 +87650,7 @@ var require_commonjs14 = __commonJS({ this.#disposed = void 0; } this.#hasDispose = !!this.#dispose; + this.#hasOnInsert = !!this.#onInsert; this.#hasDisposeAfter = !!this.#disposeAfter; this.noDisposeOnSet = !!noDisposeOnSet; this.noUpdateTTL = !!noUpdateTTL; @@ -87953,7 +88053,7 @@ var require_commonjs14 = __commonJS({ } /** * Return an array of [key, {@link LRUCache.Entry}] tuples which can be - * passed to {@link LRLUCache#load}. + * passed to {@link LRUCache#load}. * * The `start` fields are calculated relative to a portable `Date.now()` * timestamp, even if `performance.now()` is available. @@ -88064,6 +88164,9 @@ var require_commonjs14 = __commonJS({ if (status) status.set = "add"; noUpdateTTL = false; + if (this.#hasOnInsert) { + this.#onInsert?.(v, k, "add"); + } } else { this.#moveToTail(index); const oldVal = this.#valList[index]; @@ -88099,6 +88202,9 @@ var require_commonjs14 = __commonJS({ } else if (status) { status.set = "update"; } + if (this.#hasOnInsert) { + this.onInsert?.(v, k, v === oldVal ? "update" : "replace"); + } } if (ttl !== 0 && !this.#ttls) { this.#initializeTTLTracking(); @@ -88624,7 +88730,7 @@ var require_commonjs14 = __commonJS({ }); // node_modules/minipass/dist/commonjs/index.js -var require_commonjs15 = __commonJS({ +var require_commonjs17 = __commonJS({ "node_modules/minipass/dist/commonjs/index.js"(exports2) { "use strict"; var __importDefault4 = exports2 && exports2.__importDefault || function(mod) { @@ -89515,9 +89621,9 @@ var require_commonjs15 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js -var require_commonjs16 = __commonJS({ - "node_modules/archiver-utils/node_modules/path-scurry/dist/commonjs/index.js"(exports2) { +// node_modules/path-scurry/dist/commonjs/index.js +var require_commonjs18 = __commonJS({ + "node_modules/path-scurry/dist/commonjs/index.js"(exports2) { "use strict"; var __createBinding4 = exports2 && exports2.__createBinding || (Object.create ? (function(o, m, k, k2) { if (k2 === void 0) k2 = k; @@ -89548,14 +89654,14 @@ var require_commonjs16 = __commonJS({ }; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.PathScurry = exports2.Path = exports2.PathScurryDarwin = exports2.PathScurryPosix = exports2.PathScurryWin32 = exports2.PathScurryBase = exports2.PathPosix = exports2.PathWin32 = exports2.PathBase = exports2.ChildrenCache = exports2.ResolveCache = void 0; - var lru_cache_1 = require_commonjs14(); + var lru_cache_1 = require_commonjs16(); var node_path_1 = require("node:path"); var node_url_1 = require("node:url"); var fs_1 = require("fs"); var actualFS = __importStar4(require("node:fs")); var realpathSync = fs_1.realpathSync.native; var promises_1 = require("node:fs/promises"); - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var defaultFS = { lstatSync: fs_1.lstatSync, readdir: fs_1.readdir, @@ -89770,6 +89876,8 @@ var require_commonjs16 = __commonJS({ /** * Deprecated alias for Dirent['parentPath'] Somewhat counterintuitively, * this property refers to the *parent* path, not the path object itself. + * + * @deprecated */ get path() { return this.parentPath; @@ -91289,13 +91397,13 @@ var require_commonjs16 = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js +// node_modules/glob/dist/commonjs/pattern.js var require_pattern = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/pattern.js"(exports2) { + "node_modules/glob/dist/commonjs/pattern.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Pattern = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var isPatternList = (pl) => pl.length >= 1; var isGlobList = (gl) => gl.length >= 1; var Pattern = class _Pattern { @@ -91463,13 +91571,13 @@ var require_pattern = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js +// node_modules/glob/dist/commonjs/ignore.js var require_ignore = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/ignore.js"(exports2) { + "node_modules/glob/dist/commonjs/ignore.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Ignore = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var pattern_js_1 = require_pattern(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; var Ignore = class { @@ -91560,13 +91668,13 @@ var require_ignore = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js +// node_modules/glob/dist/commonjs/processor.js var require_processor = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/processor.js"(exports2) { + "node_modules/glob/dist/commonjs/processor.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Processor = exports2.SubWalks = exports2.MatchRecord = exports2.HasWalkedCache = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var HasWalkedCache = class _HasWalkedCache { store; constructor(store = /* @__PURE__ */ new Map()) { @@ -91793,13 +91901,13 @@ var require_processor = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js +// node_modules/glob/dist/commonjs/walker.js var require_walker = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/walker.js"(exports2) { + "node_modules/glob/dist/commonjs/walker.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.GlobStream = exports2.GlobWalker = exports2.GlobUtil = void 0; - var minipass_1 = require_commonjs15(); + var minipass_1 = require_commonjs17(); var ignore_js_1 = require_ignore(); var processor_js_1 = require_processor(); var makeIgnore = (ignore, opts) => typeof ignore === "string" ? new ignore_js_1.Ignore([ignore], opts) : Array.isArray(ignore) ? new ignore_js_1.Ignore(ignore, opts) : ignore; @@ -92133,15 +92241,15 @@ var require_walker = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js +// node_modules/glob/dist/commonjs/glob.js var require_glob = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/glob.js"(exports2) { + "node_modules/glob/dist/commonjs/glob.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.Glob = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var node_url_1 = require("node:url"); - var path_scurry_1 = require_commonjs16(); + var path_scurry_1 = require_commonjs18(); var pattern_js_1 = require_pattern(); var walker_js_1 = require_walker(); var defaultPlatform = typeof process === "object" && process && typeof process.platform === "string" ? process.platform : "linux"; @@ -92346,13 +92454,13 @@ var require_glob = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js +// node_modules/glob/dist/commonjs/has-magic.js var require_has_magic = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/has-magic.js"(exports2) { + "node_modules/glob/dist/commonjs/has-magic.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.hasMagic = void 0; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var hasMagic = (pattern, options = {}) => { if (!Array.isArray(pattern)) { pattern = [pattern]; @@ -92367,9 +92475,9 @@ var require_has_magic = __commonJS({ } }); -// node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js -var require_commonjs17 = __commonJS({ - "node_modules/archiver-utils/node_modules/glob/dist/commonjs/index.js"(exports2) { +// node_modules/glob/dist/commonjs/index.js +var require_commonjs19 = __commonJS({ + "node_modules/glob/dist/commonjs/index.js"(exports2) { "use strict"; Object.defineProperty(exports2, "__esModule", { value: true }); exports2.glob = exports2.sync = exports2.iterate = exports2.iterateSync = exports2.stream = exports2.streamSync = exports2.Ignore = exports2.hasMagic = exports2.Glob = exports2.unescape = exports2.escape = void 0; @@ -92378,10 +92486,10 @@ var require_commonjs17 = __commonJS({ exports2.globSync = globSync; exports2.globIterateSync = globIterateSync; exports2.globIterate = globIterate; - var minimatch_1 = require_commonjs13(); + var minimatch_1 = require_commonjs15(); var glob_js_1 = require_glob(); var has_magic_js_1 = require_has_magic(); - var minimatch_2 = require_commonjs13(); + var minimatch_2 = require_commonjs15(); Object.defineProperty(exports2, "escape", { enumerable: true, get: function() { return minimatch_2.escape; } }); @@ -92458,7 +92566,7 @@ var require_file3 = __commonJS({ var difference = require_difference(); var union = require_union(); var isPlainObject = require_isPlainObject(); - var glob2 = require_commonjs17(); + var glob2 = require_commonjs19(); var file = module2.exports = {}; var pathSeparatorRe = /[\/\\]/g; var processPatterns = function(patterns, fn) { @@ -94185,18 +94293,18 @@ var require_zip_archive_output_stream = __commonJS({ ZipArchiveOutputStream.prototype._smartStream = function(ae, callback) { var deflate = ae.getMethod() === constants.METHOD_DEFLATED; var process2 = deflate ? new DeflateCRC32Stream(this.options.zlib) : new CRC32Stream(); - var error4 = null; + var error3 = null; function handleStuff() { var digest = process2.digest().readUInt32BE(0); ae.setCrc(digest); ae.setSize(process2.size()); ae.setCompressedSize(process2.size(true)); this._afterAppend(ae); - callback(error4, ae); + callback(error3, ae); } process2.once("end", handleStuff.bind(this)); process2.once("error", function(err) { - error4 = err; + error3 = err; }); process2.pipe(this, { end: false }); return process2; @@ -95562,11 +95670,11 @@ var require_streamx = __commonJS({ } [asyncIterator]() { const stream = this; - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("readable", onreadable); this.on("close", onclose); @@ -95598,7 +95706,7 @@ var require_streamx = __commonJS({ } function ondata(data) { if (promiseReject === null) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else if (data === null && (stream._duplexState & READ_DONE) === 0) promiseReject(STREAM_DESTROYED); else promiseResolve({ value: data, done: data === null }); promiseReject = promiseResolve = null; @@ -95772,7 +95880,7 @@ var require_streamx = __commonJS({ if (all.length < 2) throw new Error("Pipeline requires at least 2 streams"); let src = all[0]; let dest = null; - let error4 = null; + let error3 = null; for (let i = 1; i < all.length; i++) { dest = all[i]; if (isStreamx(src)) { @@ -95787,14 +95895,14 @@ var require_streamx = __commonJS({ let fin = false; const autoDestroy = isStreamx(dest) || !!(dest._writableState && dest._writableState.autoDestroy); dest.on("error", (err) => { - if (error4 === null) error4 = err; + if (error3 === null) error3 = err; }); dest.on("finish", () => { fin = true; - if (!autoDestroy) done(error4); + if (!autoDestroy) done(error3); }); if (autoDestroy) { - dest.on("close", () => done(error4 || (fin ? null : PREMATURE_CLOSE))); + dest.on("close", () => done(error3 || (fin ? null : PREMATURE_CLOSE))); } } return dest; @@ -95807,8 +95915,8 @@ var require_streamx = __commonJS({ } } function onerror(err) { - if (!err || error4) return; - error4 = err; + if (!err || error3) return; + error3 = err; for (const s of all) { s.destroy(err); } @@ -96374,7 +96482,7 @@ var require_extract = __commonJS({ cb(null); } [Symbol.asyncIterator]() { - let error4 = null; + let error3 = null; let promiseResolve = null; let promiseReject = null; let entryStream = null; @@ -96382,7 +96490,7 @@ var require_extract = __commonJS({ const extract2 = this; this.on("entry", onentry); this.on("error", (err) => { - error4 = err; + error3 = err; }); this.on("close", onclose); return { @@ -96406,8 +96514,8 @@ var require_extract = __commonJS({ cb(err); } function onnext(resolve2, reject) { - if (error4) { - return reject(error4); + if (error3) { + return reject(error3); } if (entryStream) { resolve2({ value: entryStream, done: false }); @@ -96433,9 +96541,9 @@ var require_extract = __commonJS({ } } function onclose() { - consumeCallback(error4); + consumeCallback(error3); if (!promiseResolve) return; - if (error4) promiseReject(error4); + if (error3) promiseReject(error3); else promiseResolve({ value: void 0, done: true }); promiseResolve = promiseReject = null; } @@ -97331,18 +97439,18 @@ var require_zip2 = __commonJS({ return zipUploadStream; }); } - var zipErrorCallback = (error4) => { + var zipErrorCallback = (error3) => { core14.error("An error has occurred while creating the zip file for upload"); - core14.info(error4); + core14.info(error3); throw new Error("An error has occurred during zip creation for the artifact"); }; - var zipWarningCallback = (error4) => { - if (error4.code === "ENOENT") { + var zipWarningCallback = (error3) => { + if (error3.code === "ENOENT") { core14.warning("ENOENT warning during artifact zip creation. No such file or directory"); - core14.info(error4); + core14.info(error3); } else { - core14.warning(`A non-blocking warning has occurred during artifact zip creation: ${error4.code}`); - core14.info(error4); + core14.warning(`A non-blocking warning has occurred during artifact zip creation: ${error3.code}`); + core14.info(error3); } }; var zipFinishCallback = () => { @@ -99149,8 +99257,8 @@ var require_parser_stream = __commonJS({ this.unzipStream.on("entry", function(entry) { self2.push(entry); }); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(ParserStream, Transform); @@ -99290,8 +99398,8 @@ var require_extract2 = __commonJS({ this.createdDirectories = {}; var self2 = this; this.unzipStream.on("entry", this._processEntry.bind(this)); - this.unzipStream.on("error", function(error4) { - self2.emit("error", error4); + this.unzipStream.on("error", function(error3) { + self2.emit("error", error3); }); } util.inherits(Extract, Transform); @@ -99325,8 +99433,8 @@ var require_extract2 = __commonJS({ self2.unfinishedEntries--; self2._notifyAwaiter(); }); - pipedStream.on("error", function(error4) { - self2.emit("error", error4); + pipedStream.on("error", function(error3) { + self2.emit("error", error3); }); entry.pipe(pipedStream); }; @@ -99461,11 +99569,11 @@ var require_download_artifact = __commonJS({ try { yield promises_1.default.access(path2); return true; - } catch (error4) { - if (error4.code === "ENOENT") { + } catch (error3) { + if (error3.code === "ENOENT") { return false; } else { - throw error4; + throw error3; } } }); @@ -99476,9 +99584,9 @@ var require_download_artifact = __commonJS({ while (retryCount < 5) { try { return yield streamExtractExternal(url, directory); - } catch (error4) { + } catch (error3) { retryCount++; - core14.debug(`Failed to download artifact after ${retryCount} retries due to ${error4.message}. Retrying in 5 seconds...`); + core14.debug(`Failed to download artifact after ${retryCount} retries due to ${error3.message}. Retrying in 5 seconds...`); yield new Promise((resolve2) => setTimeout(resolve2, 5e3)); } } @@ -99507,10 +99615,10 @@ var require_download_artifact = __commonJS({ const extractStream = passThrough; extractStream.on("data", () => { timer.refresh(); - }).on("error", (error4) => { - core14.debug(`response.message: Artifact download failed: ${error4.message}`); + }).on("error", (error3) => { + core14.debug(`response.message: Artifact download failed: ${error3.message}`); clearTimeout(timer); - reject(error4); + reject(error3); }).pipe(unzip_stream_1.default.Extract({ path: directory })).on("close", () => { clearTimeout(timer); if (hashStream) { @@ -99519,8 +99627,8 @@ var require_download_artifact = __commonJS({ core14.info(`SHA256 digest of downloaded artifact is ${sha256Digest}`); } resolve2({ sha256Digest: `sha256:${sha256Digest}` }); - }).on("error", (error4) => { - reject(error4); + }).on("error", (error3) => { + reject(error3); }); }); }); @@ -99559,8 +99667,8 @@ var require_download_artifact = __commonJS({ core14.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -99602,8 +99710,8 @@ Are you trying to download from a different run? Try specifying a github-token w core14.debug(`Expected digest: ${options.expectedHash}`); } } - } catch (error4) { - throw new Error(`Unable to download and extract artifact: ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to download and extract artifact: ${error3.message}`); } return { downloadPath, digestMismatch }; }); @@ -99701,9 +99809,9 @@ var require_dist_node16 = __commonJS({ return request(options).then((response) => { octokit.log.info(`${requestOptions.method} ${path2} - ${response.status} in ${Date.now() - start}ms`); return response; - }).catch((error4) => { - octokit.log.info(`${requestOptions.method} ${path2} - ${error4.status} in ${Date.now() - start}ms`); - throw error4; + }).catch((error3) => { + octokit.log.info(`${requestOptions.method} ${path2} - ${error3.status} in ${Date.now() - start}ms`); + throw error3; }); }); } @@ -99721,22 +99829,22 @@ var require_dist_node17 = __commonJS({ return ex && typeof ex === "object" && "default" in ex ? ex["default"] : ex; } var Bottleneck = _interopDefault(require_light()); - async function errorRequest(octokit, state, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(octokit, state, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } async function wrapRequest(state, request, options) { const limiter = new Bottleneck(); - limiter.on("failed", function(error4, info7) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info7) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info7.retryCount + 1; if (maxRetries > info7.retryCount) { return after * state.retryAfterBaseValue; @@ -99758,12 +99866,12 @@ var require_dist_node17 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -100254,13 +100362,13 @@ var require_client2 = __commonJS({ throw new errors_1.GHESNotSupportedError(); } return (0, upload_artifact_1.uploadArtifact)(name, files, rootDirectory, options); - } catch (error4) { - (0, core_1.warning)(`Artifact upload failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Artifact upload failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions is operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -100275,13 +100383,13 @@ If the error persists, please check whether Actions is operating normally at [ht return (0, download_artifact_1.downloadArtifactPublic)(artifactId, repositoryOwner, repositoryName, token, downloadOptions); } return (0, download_artifact_1.downloadArtifactInternal)(artifactId, options); - } catch (error4) { - (0, core_1.warning)(`Download Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Download Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -100296,13 +100404,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, list_artifacts_1.listArtifactsPublic)(workflowRunId, repositoryOwner, repositoryName, token, options === null || options === void 0 ? void 0 : options.latest); } return (0, list_artifacts_1.listArtifactsInternal)(options === null || options === void 0 ? void 0 : options.latest); - } catch (error4) { - (0, core_1.warning)(`Listing Artifacts failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Listing Artifacts failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -100317,13 +100425,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, get_artifact_1.getArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, get_artifact_1.getArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Get Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Get Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -100338,13 +100446,13 @@ If the error persists, please check whether Actions and API requests are operati return (0, delete_artifact_1.deleteArtifactPublic)(artifactName, workflowRunId, repositoryOwner, repositoryName, token); } return (0, delete_artifact_1.deleteArtifactInternal)(artifactName); - } catch (error4) { - (0, core_1.warning)(`Delete Artifact failed with error: ${error4}. + } catch (error3) { + (0, core_1.warning)(`Delete Artifact failed with error: ${error3}. Errors can be temporary, so please try again and optionally run the action with debug mode enabled for more information. If the error persists, please check whether Actions and API requests are operating normally at [https://githubstatus.com](https://www.githubstatus.com).`); - throw error4; + throw error3; } }); } @@ -100844,14 +100952,14 @@ var require_tmp = __commonJS({ options.template = _getRelativePathSync("template", options.template, tmpDir); return options; } - function _isEBADF(error4) { - return _isExpectedError(error4, -EBADF, "EBADF"); + function _isEBADF(error3) { + return _isExpectedError(error3, -EBADF, "EBADF"); } - function _isENOENT(error4) { - return _isExpectedError(error4, -ENOENT, "ENOENT"); + function _isENOENT(error3) { + return _isExpectedError(error3, -ENOENT, "ENOENT"); } - function _isExpectedError(error4, errno, code) { - return IS_WIN32 ? error4.code === code : error4.code === code && error4.errno === errno; + function _isExpectedError(error3, errno, code) { + return IS_WIN32 ? error3.code === code : error3.code === code && error3.errno === errno; } function setGracefulCleanup() { _gracefulCleanup = true; @@ -102563,9 +102671,9 @@ var require_upload_gzip = __commonJS({ const size = (yield stat(tempFilePath)).size; resolve2(size); })); - outputStream.on("error", (error4) => { - console.log(error4); - reject(error4); + outputStream.on("error", (error3) => { + console.log(error3); + reject(error3); }); }); }); @@ -102690,9 +102798,9 @@ var require_requestUtils = __commonJS({ } isRetryable = (0, utils_1.isRetryableStatusCode)(statusCode); errorMessage = `Artifact service responded with ${statusCode}`; - } catch (error4) { + } catch (error3) { isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { core14.info(`${name} - Error is not retryable`); @@ -103058,9 +103166,9 @@ var require_upload_http_client = __commonJS({ let response; try { response = yield uploadChunkRequest(); - } catch (error4) { + } catch (error3) { core14.info(`An error has been caught http-client index ${httpClientIndex}, retrying the upload`); - console.log(error4); + console.log(error3); if (incrementAndCheckRetryLimit()) { return false; } @@ -103249,8 +103357,8 @@ var require_download_http_client = __commonJS({ } this.statusReporter.incrementProcessedCount(); } - }))).catch((error4) => { - throw new Error(`Unable to download the artifact: ${error4}`); + }))).catch((error3) => { + throw new Error(`Unable to download the artifact: ${error3}`); }).finally(() => { this.statusReporter.stop(); this.downloadHttpManager.disposeAndReplaceAllClients(); @@ -103315,9 +103423,9 @@ var require_download_http_client = __commonJS({ let response; try { response = yield makeDownloadRequest(); - } catch (error4) { + } catch (error3) { core14.info("An error occurred while attempting to download a file"); - console.log(error4); + console.log(error3); yield backOff(); continue; } @@ -103331,7 +103439,7 @@ var require_download_http_client = __commonJS({ } else { forceRetry = true; } - } catch (error4) { + } catch (error3) { forceRetry = true; } } @@ -103357,31 +103465,31 @@ var require_download_http_client = __commonJS({ yield new Promise((resolve2, reject) => { if (isGzip) { const gunzip = zlib.createGunzip(); - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core14.info(`An error occurred while attempting to read the response stream`); gunzip.close(); destinationStream.close(); - reject(error4); - }).pipe(gunzip).on("error", (error4) => { + reject(error3); + }).pipe(gunzip).on("error", (error3) => { core14.info(`An error occurred while attempting to decompress the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve2(); - }).on("error", (error4) => { + }).on("error", (error3) => { core14.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } else { - response.message.on("error", (error4) => { + response.message.on("error", (error3) => { core14.info(`An error occurred while attempting to read the response stream`); destinationStream.close(); - reject(error4); + reject(error3); }).pipe(destinationStream).on("close", () => { resolve2(); - }).on("error", (error4) => { + }).on("error", (error3) => { core14.info(`An error occurred while writing a downloaded file to ${destinationStream.path}`); - reject(error4); + reject(error3); }); } }); @@ -105273,7 +105381,7 @@ var require_concat_map = __commonJS({ }); // node_modules/brace-expansion/index.js -var require_brace_expansion3 = __commonJS({ +var require_brace_expansion2 = __commonJS({ "node_modules/brace-expansion/index.js"(exports2, module2) { var concatMap = require_concat_map(); var balanced = require_balanced_match(); @@ -105431,7 +105539,7 @@ var require_minimatch2 = __commonJS({ }; minimatch.sep = path2.sep; var GLOBSTAR = minimatch.GLOBSTAR = Minimatch.GLOBSTAR = {}; - var expand = require_brace_expansion3(); + var expand = require_brace_expansion2(); var plTypes = { "!": { open: "(?:(?!(?:", close: "))[^/]*?)" }, "?": { open: "(?:", close: ")?" }, @@ -108299,7 +108407,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core14 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob2 = __importStar4(require_glob2()); var io6 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -108383,7 +108491,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core14.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -109501,9 +109609,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -109617,12 +109725,12 @@ var require_requestUtils2 = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -109656,13 +109764,13 @@ var require_requestUtils2 = __commonJS({ delay, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -110478,8 +110586,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -111241,8 +111349,8 @@ var require_util15 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -111338,8 +111446,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -111370,18 +111478,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -111649,8 +111757,8 @@ var require_tar2 = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -111851,22 +111959,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -111921,15 +112029,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -111937,8 +112045,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -112000,10 +112108,10 @@ var require_cache3 = __commonJS({ } core14.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else { @@ -112016,8 +112124,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -112062,8 +112170,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core14.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core14.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError2(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core14.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -112082,10 +112190,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError2.name) { core14.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -112100,8 +112208,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -114827,7 +114935,7 @@ var require_debug2 = __commonJS({ if (!debug4) { try { debug4 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug4 !== "function") { debug4 = function() { @@ -114860,8 +114968,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -114935,9 +115043,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -115104,10 +115212,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -115306,12 +115414,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -116443,7 +116551,6 @@ var io2 = __toESM(require_io2()); // src/util.ts var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js @@ -116461,14 +116568,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs2.lstat(itemPath, { bigint: true }) : await fs2.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs2.lstat(itemPath, { bigint: true }) : await fs2.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs2.readdir(itemPath) : await fs2.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs2.readdir(itemPath) : await fs2.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -116479,13 +116586,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -117365,6 +117472,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -117484,7 +117603,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -117524,16 +117643,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -119167,8 +119277,8 @@ var ConfigurationError = class extends Error { super(message); } }; -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } // src/actions-util.ts @@ -119466,7 +119576,7 @@ function withGroup(groupName, f) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; @@ -119538,6 +119648,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -119649,21 +119764,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, @@ -119676,6 +119791,8 @@ var featureConfig = { var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, @@ -119846,9 +119963,9 @@ async function runWrapper() { ) ); } - } catch (error4) { + } catch (error3) { core13.setFailed( - `upload-sarif post-action step failed: ${getErrorMessage(error4)}` + `upload-sarif post-action step failed: ${getErrorMessage(error3)}` ); } } @@ -119937,5 +120054,5 @@ tmp/lib/tmp.js: *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) */ diff --git a/lib/upload-sarif-action.js b/lib/upload-sarif-action.js index 1266953f0..9ee03c88c 100644 --- a/lib/upload-sarif-action.js +++ b/lib/upload-sarif-action.js @@ -426,18 +426,18 @@ var require_tunnel = __commonJS({ res.statusCode ); socket.destroy(); - var error4 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, statusCode=" + res.statusCode); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } if (head.length > 0) { debug6("got illegal response body from proxy"); socket.destroy(); - var error4 = new Error("got illegal response body from proxy"); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("got illegal response body from proxy"); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); return; } @@ -452,9 +452,9 @@ var require_tunnel = __commonJS({ cause.message, cause.stack ); - var error4 = new Error("tunneling socket could not be established, cause=" + cause.message); - error4.code = "ECONNRESET"; - options.request.emit("error", error4); + var error3 = new Error("tunneling socket could not be established, cause=" + cause.message); + error3.code = "ECONNRESET"; + options.request.emit("error", error3); self2.removeSocket(placeholder); } }; @@ -5582,7 +5582,7 @@ Content-Type: ${value.type || "application/octet-stream"}\r throw new TypeError("Body is unusable"); } const promise = createDeferredPromise(); - const errorSteps = (error4) => promise.reject(error4); + const errorSteps = (error3) => promise.reject(error3); const successSteps = (data) => { try { promise.resolve(convertBytesToJSValue(data)); @@ -5868,16 +5868,16 @@ var require_request = __commonJS({ this.onError(err); } } - onError(error4) { + onError(error3) { this.onFinally(); if (channels.error.hasSubscribers) { - channels.error.publish({ request: this, error: error4 }); + channels.error.publish({ request: this, error: error3 }); } if (this.aborted) { return; } this.aborted = true; - return this[kHandler].onError(error4); + return this[kHandler].onError(error3); } onFinally() { if (this.errorHandler) { @@ -6740,8 +6740,8 @@ var require_RedirectHandler = __commonJS({ onUpgrade(statusCode, headers, socket) { this.handler.onUpgrade(statusCode, headers, socket); } - onError(error4) { - this.handler.onError(error4); + onError(error3) { + this.handler.onError(error3); } onHeaders(statusCode, headers, resume, statusText) { this.location = this.history.length >= this.maxRedirections || util.isDisturbed(this.opts.body) ? null : parseLocation(statusCode, headers); @@ -8882,7 +8882,7 @@ var require_pool = __commonJS({ this[kOptions] = { ...util.deepClone(options), connect, allowH2 }; this[kOptions].interceptors = options.interceptors ? { ...options.interceptors } : void 0; this[kFactory] = factory; - this.on("connectionError", (origin2, targets, error4) => { + this.on("connectionError", (origin2, targets, error3) => { for (const target of targets) { const idx = this[kClients].indexOf(target); if (idx !== -1) { @@ -10491,13 +10491,13 @@ var require_mock_utils = __commonJS({ if (mockDispatch2.data.callback) { mockDispatch2.data = { ...mockDispatch2.data, ...mockDispatch2.data.callback(opts) }; } - const { data: { statusCode, data, headers, trailers, error: error4 }, delay: delay2, persist } = mockDispatch2; + const { data: { statusCode, data, headers, trailers, error: error3 }, delay: delay2, persist } = mockDispatch2; const { timesInvoked, times } = mockDispatch2; mockDispatch2.consumed = !persist && timesInvoked >= times; mockDispatch2.pending = timesInvoked < times; - if (error4 !== null) { + if (error3 !== null) { deleteMockDispatch(this[kDispatches], key); - handler.onError(error4); + handler.onError(error3); return true; } if (typeof delay2 === "number" && delay2 > 0) { @@ -10535,19 +10535,19 @@ var require_mock_utils = __commonJS({ if (agent.isMockActive) { try { mockDispatch.call(this, opts, handler); - } catch (error4) { - if (error4 instanceof MockNotMatchedError) { + } catch (error3) { + if (error3 instanceof MockNotMatchedError) { const netConnect = agent[kGetNetConnect](); if (netConnect === false) { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect disabled)`); } if (checkNetConnect(netConnect, origin)) { originalDispatch.call(this, opts, handler); } else { - throw new MockNotMatchedError(`${error4.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); + throw new MockNotMatchedError(`${error3.message}: subsequent request to origin ${origin} was not allowed (net.connect is not enabled for this origin)`); } } else { - throw error4; + throw error3; } } } else { @@ -10710,11 +10710,11 @@ var require_mock_interceptor = __commonJS({ /** * Mock an undici request with a defined error. */ - replyWithError(error4) { - if (typeof error4 === "undefined") { + replyWithError(error3) { + if (typeof error3 === "undefined") { throw new InvalidArgumentError("error must be defined"); } - const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error4 }); + const newMockDispatch = addMockDispatch(this[kDispatches], this[kDispatchKey], { error: error3 }); return new MockScope(newMockDispatch); } /** @@ -13041,17 +13041,17 @@ var require_fetch = __commonJS({ this.emit("terminated", reason); } // https://fetch.spec.whatwg.org/#fetch-controller-abort - abort(error4) { + abort(error3) { if (this.state !== "ongoing") { return; } this.state = "aborted"; - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - this.serializedAbortReason = error4; - this.connection?.destroy(error4); - this.emit("terminated", error4); + this.serializedAbortReason = error3; + this.connection?.destroy(error3); + this.emit("terminated", error3); } }; function fetch(input, init = {}) { @@ -13155,13 +13155,13 @@ var require_fetch = __commonJS({ performance.markResourceTiming(timingInfo, originalURL.href, initiatorType, globalThis2, cacheState); } } - function abortFetch(p, request, responseObject, error4) { - if (!error4) { - error4 = new DOMException2("The operation was aborted.", "AbortError"); + function abortFetch(p, request, responseObject, error3) { + if (!error3) { + error3 = new DOMException2("The operation was aborted.", "AbortError"); } - p.reject(error4); + p.reject(error3); if (request.body != null && isReadable(request.body?.stream)) { - request.body.stream.cancel(error4).catch((err) => { + request.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13173,7 +13173,7 @@ var require_fetch = __commonJS({ } const response = responseObject[kState]; if (response.body != null && isReadable(response.body?.stream)) { - response.body.stream.cancel(error4).catch((err) => { + response.body.stream.cancel(error3).catch((err) => { if (err.code === "ERR_INVALID_STATE") { return; } @@ -13953,13 +13953,13 @@ var require_fetch = __commonJS({ fetchParams.controller.ended = true; this.body.push(null); }, - onError(error4) { + onError(error3) { if (this.abort) { fetchParams.controller.off("terminated", this.abort); } - this.body?.destroy(error4); - fetchParams.controller.terminate(error4); - reject(error4); + this.body?.destroy(error3); + fetchParams.controller.terminate(error3); + reject(error3); }, onUpgrade(status, headersList, socket) { if (status !== 101) { @@ -14425,8 +14425,8 @@ var require_util4 = __commonJS({ } fr[kResult] = result; fireAProgressEvent("load", fr); - } catch (error4) { - fr[kError] = error4; + } catch (error3) { + fr[kError] = error3; fireAProgressEvent("error", fr); } if (fr[kState] !== "loading") { @@ -14435,13 +14435,13 @@ var require_util4 = __commonJS({ }); break; } - } catch (error4) { + } catch (error3) { if (fr[kAborted]) { return; } queueMicrotask(() => { fr[kState] = "done"; - fr[kError] = error4; + fr[kError] = error3; fireAProgressEvent("error", fr); if (fr[kState] !== "loading") { fireAProgressEvent("loadend", fr); @@ -16441,11 +16441,11 @@ var require_connection = __commonJS({ }); } } - function onSocketError(error4) { + function onSocketError(error3) { const { ws } = this; ws[kReadyState] = states.CLOSING; if (channels.socketError.hasSubscribers) { - channels.socketError.publish(error4); + channels.socketError.publish(error3); } this.destroy(); } @@ -18077,12 +18077,12 @@ var require_oidc_utils = __commonJS({ var _a; return __awaiter4(this, void 0, void 0, function* () { const httpclient = _OidcClient.createHttpClient(); - const res = yield httpclient.getJson(id_token_url).catch((error4) => { + const res = yield httpclient.getJson(id_token_url).catch((error3) => { throw new Error(`Failed to get ID Token. - Error Code : ${error4.statusCode} + Error Code : ${error3.statusCode} - Error Message: ${error4.message}`); + Error Message: ${error3.message}`); }); const id_token = (_a = res.result) === null || _a === void 0 ? void 0 : _a.value; if (!id_token) { @@ -18103,8 +18103,8 @@ var require_oidc_utils = __commonJS({ const id_token = yield _OidcClient.getCall(id_token_url); (0, core_1.setSecret)(id_token); return id_token; - } catch (error4) { - throw new Error(`Error message: ${error4.message}`); + } catch (error3) { + throw new Error(`Error message: ${error3.message}`); } }); } @@ -19226,7 +19226,7 @@ var require_toolrunner = __commonJS({ this._debug(`STDIO streams have closed for tool '${this.toolPath}'`); state.CheckComplete(); }); - state.on("done", (error4, exitCode) => { + state.on("done", (error3, exitCode) => { if (stdbuffer.length > 0) { this.emit("stdline", stdbuffer); } @@ -19234,8 +19234,8 @@ var require_toolrunner = __commonJS({ this.emit("errline", errbuffer); } cp.removeAllListeners(); - if (error4) { - reject(error4); + if (error3) { + reject(error3); } else { resolve6(exitCode); } @@ -19330,14 +19330,14 @@ var require_toolrunner = __commonJS({ this.emit("debug", message); } _setResult() { - let error4; + let error3; if (this.processExited) { if (this.processError) { - error4 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); + error3 = new Error(`There was an error when attempting to execute the process '${this.toolPath}'. This may indicate the process failed to start. Error: ${this.processError}`); } else if (this.processExitCode !== 0 && !this.options.ignoreReturnCode) { - error4 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); + error3 = new Error(`The process '${this.toolPath}' failed with exit code ${this.processExitCode}`); } else if (this.processStderr && this.options.failOnStdErr) { - error4 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); + error3 = new Error(`The process '${this.toolPath}' failed because one or more lines were written to the STDERR stream`); } } if (this.timeout) { @@ -19345,7 +19345,7 @@ var require_toolrunner = __commonJS({ this.timeout = null; } this.done = true; - this.emit("done", error4, this.processExitCode); + this.emit("done", error3, this.processExitCode); } static HandleTimeout(state) { if (state.done) { @@ -19419,7 +19419,7 @@ var require_exec = __commonJS({ exports2.getExecOutput = exports2.exec = void 0; var string_decoder_1 = require("string_decoder"); var tr = __importStar4(require_toolrunner()); - function exec2(commandLine, args, options) { + function exec(commandLine, args, options) { return __awaiter4(this, void 0, void 0, function* () { const commandArgs = tr.argStringToArray(commandLine); if (commandArgs.length === 0) { @@ -19431,8 +19431,8 @@ var require_exec = __commonJS({ return runner.exec(); }); } - exports2.exec = exec2; - function getExecOutput2(commandLine, args, options) { + exports2.exec = exec; + function getExecOutput(commandLine, args, options) { var _a, _b; return __awaiter4(this, void 0, void 0, function* () { let stdout = ""; @@ -19454,7 +19454,7 @@ var require_exec = __commonJS({ } }; const listeners = Object.assign(Object.assign({}, options === null || options === void 0 ? void 0 : options.listeners), { stdout: stdOutListener, stderr: stdErrListener }); - const exitCode = yield exec2(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); + const exitCode = yield exec(commandLine, args, Object.assign(Object.assign({}, options), { listeners })); stdout += stdoutDecoder.end(); stderr += stderrDecoder.end(); return { @@ -19464,7 +19464,7 @@ var require_exec = __commonJS({ }; }); } - exports2.getExecOutput = getExecOutput2; + exports2.getExecOutput = getExecOutput; } }); @@ -19532,12 +19532,12 @@ var require_platform = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getDetails = exports2.isLinux = exports2.isMacOS = exports2.isWindows = exports2.arch = exports2.platform = void 0; var os_1 = __importDefault4(require("os")); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var getWindowsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout: version } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { + const { stdout: version } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Version"', void 0, { silent: true }); - const { stdout: name } = yield exec2.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { + const { stdout: name } = yield exec.getExecOutput('powershell -command "(Get-CimInstance -ClassName Win32_OperatingSystem).Caption"', void 0, { silent: true }); return { @@ -19547,7 +19547,7 @@ var require_platform = __commonJS({ }); var getMacOsInfo = () => __awaiter4(void 0, void 0, void 0, function* () { var _a, _b, _c, _d; - const { stdout } = yield exec2.getExecOutput("sw_vers", void 0, { + const { stdout } = yield exec.getExecOutput("sw_vers", void 0, { silent: true }); const version = (_b = (_a = stdout.match(/ProductVersion:\s*(.+)/)) === null || _a === void 0 ? void 0 : _a[1]) !== null && _b !== void 0 ? _b : ""; @@ -19558,7 +19558,7 @@ var require_platform = __commonJS({ }; }); var getLinuxInfo = () => __awaiter4(void 0, void 0, void 0, function* () { - const { stdout } = yield exec2.getExecOutput("lsb_release", ["-i", "-r", "-s"], { + const { stdout } = yield exec.getExecOutput("lsb_release", ["-i", "-r", "-s"], { silent: true }); const [name, version] = stdout.trim().split("\n"); @@ -19728,7 +19728,7 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); exports2.setCommandEcho = setCommandEcho; function setFailed2(message) { process.exitCode = ExitCode.Failure; - error4(message); + error3(message); } exports2.setFailed = setFailed2; function isDebug2() { @@ -19739,14 +19739,14 @@ Support boolean input list: \`true | True | TRUE | false | False | FALSE\``); (0, command_1.issueCommand)("debug", {}, message); } exports2.debug = debug6; - function error4(message, properties = {}) { + function error3(message, properties = {}) { (0, command_1.issueCommand)("error", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.error = error4; - function warning9(message, properties = {}) { + exports2.error = error3; + function warning10(message, properties = {}) { (0, command_1.issueCommand)("warning", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } - exports2.warning = warning9; + exports2.warning = warning10; function notice(message, properties = {}) { (0, command_1.issueCommand)("notice", (0, utils_1.toCommandProperties)(properties), message instanceof Error ? message.toString() : message); } @@ -20745,8 +20745,8 @@ var require_add = __commonJS({ } if (kind === "error") { hook = function(method, options) { - return Promise.resolve().then(method.bind(null, options)).catch(function(error4) { - return orig(error4, options); + return Promise.resolve().then(method.bind(null, options)).catch(function(error3) { + return orig(error3, options); }); }; } @@ -21478,7 +21478,7 @@ var require_dist_node5 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -21487,7 +21487,7 @@ var require_dist_node5 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -21497,17 +21497,17 @@ var require_dist_node5 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -22126,7 +22126,7 @@ var require_dist_node8 = __commonJS({ } if (status >= 400) { const data = await getResponseData(response); - const error4 = new import_request_error.RequestError(toErrorMessage(data), status, { + const error3 = new import_request_error.RequestError(toErrorMessage(data), status, { response: { url: url2, status, @@ -22135,7 +22135,7 @@ var require_dist_node8 = __commonJS({ }, request: requestOptions }); - throw error4; + throw error3; } return parseSuccessResponseBody ? await getResponseData(response) : response.body; }).then((data) => { @@ -22145,17 +22145,17 @@ var require_dist_node8 = __commonJS({ headers, data }; - }).catch((error4) => { - if (error4 instanceof import_request_error.RequestError) - throw error4; - else if (error4.name === "AbortError") - throw error4; - let message = error4.message; - if (error4.name === "TypeError" && "cause" in error4) { - if (error4.cause instanceof Error) { - message = error4.cause.message; - } else if (typeof error4.cause === "string") { - message = error4.cause; + }).catch((error3) => { + if (error3 instanceof import_request_error.RequestError) + throw error3; + else if (error3.name === "AbortError") + throw error3; + let message = error3.message; + if (error3.name === "TypeError" && "cause" in error3) { + if (error3.cause instanceof Error) { + message = error3.cause.message; + } else if (typeof error3.cause === "string") { + message = error3.cause; } } throw new import_request_error.RequestError(message, 500, { @@ -24827,9 +24827,9 @@ var require_dist_node13 = __commonJS({ /<([^<>]+)>;\s*rel="next"/ ) || [])[1]; return { value: normalizedResponse }; - } catch (error4) { - if (error4.status !== 409) - throw error4; + } catch (error3) { + if (error3.status !== 409) + throw error3; url2 = ""; return { value: { @@ -27627,7 +27627,7 @@ var require_package = __commonJS({ "package.json"(exports2, module2) { module2.exports = { name: "codeql", - version: "4.31.3", + version: "4.31.6", private: true, description: "CodeQL action", scripts: { @@ -27662,23 +27662,21 @@ var require_package = __commonJS({ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", archiver: "^7.0.1", "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", jsonschema: "1.4.1", long: "^5.3.2", "node-forge": "^1.3.1", - octokit: "^5.0.5", semver: "^7.7.3", uuid: "^13.0.0" }, devDependencies: { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -27686,10 +27684,10 @@ var require_package = __commonJS({ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", ava: "^6.4.1", @@ -27699,9 +27697,9 @@ var require_package = __commonJS({ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", - glob: "^11.0.3", + glob: "^11.1.0", nock: "^14.0.10", sinon: "^21.0.0", typescript: "^5.9.3" @@ -27725,7 +27723,8 @@ var require_package = __commonJS({ "eslint-plugin-jsx-a11y": { semver: ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + glob: "^11.1.0" } }; } @@ -27911,8 +27910,8 @@ var require_light = __commonJS({ } else { return returned; } - } catch (error4) { - e2 = error4; + } catch (error3) { + e2 = error3; { this.trigger("error", e2); } @@ -27922,8 +27921,8 @@ var require_light = __commonJS({ return (await Promise.all(promises3)).find(function(x) { return x != null; }); - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; { this.trigger("error", e); } @@ -28035,10 +28034,10 @@ var require_light = __commonJS({ _randomIndex() { return Math.random().toString(36).slice(2); } - doDrop({ error: error4, message = "This job has been dropped by Bottleneck" } = {}) { + doDrop({ error: error3, message = "This job has been dropped by Bottleneck" } = {}) { if (this._states.remove(this.options.id)) { if (this.rejectOnDrop) { - this._reject(error4 != null ? error4 : new BottleneckError$1(message)); + this._reject(error3 != null ? error3 : new BottleneckError$1(message)); } this.Events.trigger("dropped", { args: this.args, options: this.options, task: this.task, promise: this.promise }); return true; @@ -28072,7 +28071,7 @@ var require_light = __commonJS({ return this.Events.trigger("scheduled", { args: this.args, options: this.options }); } async doExecute(chained, clearGlobalState, run2, free) { - var error4, eventInfo, passed; + var error3, eventInfo, passed; if (this.retryCount === 0) { this._assertStatus("RUNNING"); this._states.next(this.options.id); @@ -28090,24 +28089,24 @@ var require_light = __commonJS({ return this._resolve(passed); } } catch (error1) { - error4 = error1; - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = error1; + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } } doExpire(clearGlobalState, run2, free) { - var error4, eventInfo; + var error3, eventInfo; if (this._states.jobStatus(this.options.id === "RUNNING")) { this._states.next(this.options.id); } this._assertStatus("EXECUTING"); eventInfo = { args: this.args, options: this.options, retryCount: this.retryCount }; - error4 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); - return this._onFailure(error4, eventInfo, clearGlobalState, run2, free); + error3 = new BottleneckError$1(`This job timed out after ${this.options.expiration} ms.`); + return this._onFailure(error3, eventInfo, clearGlobalState, run2, free); } - async _onFailure(error4, eventInfo, clearGlobalState, run2, free) { + async _onFailure(error3, eventInfo, clearGlobalState, run2, free) { var retry3, retryAfter; if (clearGlobalState()) { - retry3 = await this.Events.trigger("failed", error4, eventInfo); + retry3 = await this.Events.trigger("failed", error3, eventInfo); if (retry3 != null) { retryAfter = ~~retry3; this.Events.trigger("retry", `Retrying ${this.options.id} after ${retryAfter} ms`, eventInfo); @@ -28117,7 +28116,7 @@ var require_light = __commonJS({ this.doDone(eventInfo); await free(this.options, eventInfo); this._assertStatus("DONE"); - return this._reject(error4); + return this._reject(error3); } } } @@ -28396,7 +28395,7 @@ var require_light = __commonJS({ return this._queue.length === 0; } async _tryToRun() { - var args, cb, error4, reject, resolve6, returned, task; + var args, cb, error3, reject, resolve6, returned, task; if (this._running < 1 && this._queue.length > 0) { this._running++; ({ task, args, resolve: resolve6, reject } = this._queue.shift()); @@ -28407,9 +28406,9 @@ var require_light = __commonJS({ return resolve6(returned); }; } catch (error1) { - error4 = error1; + error3 = error1; return function() { - return reject(error4); + return reject(error3); }; } })(); @@ -28543,8 +28542,8 @@ var require_light = __commonJS({ } else { results.push(void 0); } - } catch (error4) { - e = error4; + } catch (error3) { + e = error3; results.push(v.Events.trigger("error", e)); } } @@ -28877,14 +28876,14 @@ var require_light = __commonJS({ return done; } async _addToQueue(job) { - var args, blocked, error4, options, reachedHWM, shifted, strategy; + var args, blocked, error3, options, reachedHWM, shifted, strategy; ({ args, options } = job); try { ({ reachedHWM, blocked, strategy } = await this._store.__submit__(this.queued(), options.weight)); } catch (error1) { - error4 = error1; - this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error4 }); - job.doDrop({ error: error4 }); + error3 = error1; + this.Events.trigger("debug", `Could not queue ${options.id}`, { args, options, error: error3 }); + job.doDrop({ error: error3 }); return false; } if (blocked) { @@ -29180,24 +29179,24 @@ var require_dist_node15 = __commonJS({ }); module2.exports = __toCommonJS2(dist_src_exports); var import_core = require_dist_node11(); - async function errorRequest(state, octokit, error4, options) { - if (!error4.request || !error4.request.request) { - throw error4; + async function errorRequest(state, octokit, error3, options) { + if (!error3.request || !error3.request.request) { + throw error3; } - if (error4.status >= 400 && !state.doNotRetry.includes(error4.status)) { + if (error3.status >= 400 && !state.doNotRetry.includes(error3.status)) { const retries = options.request.retries != null ? options.request.retries : state.retries; const retryAfter = Math.pow((options.request.retryCount || 0) + 1, 2); - throw octokit.retry.retryRequest(error4, retries, retryAfter); + throw octokit.retry.retryRequest(error3, retries, retryAfter); } - throw error4; + throw error3; } var import_light = __toESM2(require_light()); var import_request_error = require_dist_node14(); async function wrapRequest(state, octokit, request, options) { const limiter = new import_light.default(); - limiter.on("failed", function(error4, info6) { - const maxRetries = ~~error4.request.request.retries; - const after = ~~error4.request.request.retryAfter; + limiter.on("failed", function(error3, info6) { + const maxRetries = ~~error3.request.request.retries; + const after = ~~error3.request.request.retryAfter; options.request.retryCount = info6.retryCount + 1; if (maxRetries > info6.retryCount) { return after * state.retryAfterBaseValue; @@ -29213,11 +29212,11 @@ var require_dist_node15 = __commonJS({ if (response.data && response.data.errors && response.data.errors.length > 0 && /Something went wrong while executing your query/.test( response.data.errors[0].message )) { - const error4 = new import_request_error.RequestError(response.data.errors[0].message, 500, { + const error3 = new import_request_error.RequestError(response.data.errors[0].message, 500, { request: options, response }); - return errorRequest(state, octokit, error4, options); + return errorRequest(state, octokit, error3, options); } return response; } @@ -29238,12 +29237,12 @@ var require_dist_node15 = __commonJS({ } return { retry: { - retryRequest: (error4, retries, retryAfter) => { - error4.request.request = Object.assign({}, error4.request.request, { + retryRequest: (error3, retries, retryAfter) => { + error3.request.request = Object.assign({}, error3.request.request, { retries, retryAfter }); - return error4; + return error3; } } }; @@ -32638,7 +32637,7 @@ var require_cacheUtils = __commonJS({ Object.defineProperty(exports2, "__esModule", { value: true }); exports2.getRuntimeToken = exports2.getCacheVersion = exports2.assertDefined = exports2.getGnuTarPathOnWindows = exports2.getCacheFileName = exports2.getCompressionMethod = exports2.unlinkFile = exports2.resolvePaths = exports2.getArchiveFileSizeInBytes = exports2.createTempDirectory = void 0; var core14 = __importStar4(require_core()); - var exec2 = __importStar4(require_exec()); + var exec = __importStar4(require_exec()); var glob = __importStar4(require_glob()); var io6 = __importStar4(require_io3()); var crypto = __importStar4(require("crypto")); @@ -32722,7 +32721,7 @@ var require_cacheUtils = __commonJS({ additionalArgs.push("--version"); core14.debug(`Checking ${app} ${additionalArgs.join(" ")}`); try { - yield exec2.exec(`${app}`, additionalArgs, { + yield exec.exec(`${app}`, additionalArgs, { ignoreReturnCode: true, silent: true, listeners: { @@ -34788,8 +34787,8 @@ function __read(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -35023,9 +35022,9 @@ var init_tslib_es6 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default = { __extends, @@ -36356,14 +36355,14 @@ var require_browser = __commonJS({ } else { exports2.storage.removeItem("debug"); } - } catch (error4) { + } catch (error3) { } } function load2() { let r; try { r = exports2.storage.getItem("debug") || exports2.storage.getItem("DEBUG"); - } catch (error4) { + } catch (error3) { } if (!r && typeof process !== "undefined" && "env" in process) { r = process.env.DEBUG; @@ -36373,7 +36372,7 @@ var require_browser = __commonJS({ function localstorage() { try { return localStorage; - } catch (error4) { + } catch (error3) { } } module2.exports = require_common()(exports2); @@ -36381,8 +36380,8 @@ var require_browser = __commonJS({ formatters.j = function(v) { try { return JSON.stringify(v); - } catch (error4) { - return "[UnexpectedJSONParseError]: " + error4.message; + } catch (error3) { + return "[UnexpectedJSONParseError]: " + error3.message; } }; } @@ -36602,7 +36601,7 @@ var require_node = __commonJS({ 221 ]; } - } catch (error4) { + } catch (error3) { } exports2.inspectOpts = Object.keys(process.env).filter((key) => { return /^debug_/i.test(key); @@ -37812,14 +37811,14 @@ var require_tracingPolicy = __commonJS({ return void 0; } } - function tryProcessError(span, error4) { + function tryProcessError(span, error3) { try { span.setStatus({ status: "error", - error: (0, core_util_1.isError)(error4) ? error4 : void 0 + error: (0, core_util_1.isError)(error3) ? error3 : void 0 }); - if ((0, restError_js_1.isRestError)(error4) && error4.statusCode) { - span.setAttribute("http.status_code", error4.statusCode); + if ((0, restError_js_1.isRestError)(error3) && error3.statusCode) { + span.setAttribute("http.status_code", error3.statusCode); } span.end(); } catch (e) { @@ -38491,11 +38490,11 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ logger }); let response; - let error4; + let error3; try { response = await next(request); } catch (err) { - error4 = err; + error3 = err; response = err.response; } if (callbacks.authorizeRequestOnChallenge && (response === null || response === void 0 ? void 0 : response.status) === 401 && getChallenge(response)) { @@ -38510,8 +38509,8 @@ var require_bearerTokenAuthenticationPolicy = __commonJS({ return next(request); } } - if (error4) { - throw error4; + if (error3) { + throw error3; } else { return response; } @@ -39005,8 +39004,8 @@ function __read2(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -39240,9 +39239,9 @@ var init_tslib_es62 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError2 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default2 = { __extends: __extends2, @@ -39742,8 +39741,8 @@ function __read3(o, n) { var i = m.call(o), r, ar = [], e; try { while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value); - } catch (error4) { - e = { error: error4 }; + } catch (error3) { + e = { error: error3 }; } finally { try { if (r && !r.done && (m = i["return"])) m.call(i); @@ -39977,9 +39976,9 @@ var init_tslib_es63 = __esm({ }) : function(o, v) { o["default"] = v; }; - _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error4, suppressed, message) { + _SuppressedError3 = typeof SuppressedError === "function" ? SuppressedError : function(error3, suppressed, message) { var e = new Error(message); - return e.name = "SuppressedError", e.error = error4, e.suppressed = suppressed, e; + return e.name = "SuppressedError", e.error = error3, e.suppressed = suppressed, e; }; tslib_es6_default3 = { __extends: __extends3, @@ -41042,9 +41041,9 @@ var require_deserializationPolicy = __commonJS({ return parsedResponse; } const responseSpec = getOperationResponseMap(parsedResponse); - const { error: error4, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); - if (error4) { - throw error4; + const { error: error3, shouldReturnResponse } = handleErrorResponse(parsedResponse, operationSpec, responseSpec, options); + if (error3) { + throw error3; } else if (shouldReturnResponse) { return parsedResponse; } @@ -41092,13 +41091,13 @@ var require_deserializationPolicy = __commonJS({ } const errorResponseSpec = responseSpec !== null && responseSpec !== void 0 ? responseSpec : operationSpec.responses.default; const initialErrorMessage = ((_a = parsedResponse.request.streamResponseStatusCodes) === null || _a === void 0 ? void 0 : _a.has(parsedResponse.status)) ? `Unexpected status code: ${parsedResponse.status}` : parsedResponse.bodyAsText; - const error4 = new core_rest_pipeline_1.RestError(initialErrorMessage, { + const error3 = new core_rest_pipeline_1.RestError(initialErrorMessage, { statusCode: parsedResponse.status, request: parsedResponse.request, response: parsedResponse }); if (!errorResponseSpec) { - throw error4; + throw error3; } const defaultBodyMapper = errorResponseSpec.bodyMapper; const defaultHeadersMapper = errorResponseSpec.headersMapper; @@ -41118,21 +41117,21 @@ var require_deserializationPolicy = __commonJS({ deserializedError = operationSpec.serializer.deserialize(defaultBodyMapper, valueToDeserialize, "error.response.parsedBody", options); } const internalError = parsedBody.error || deserializedError || parsedBody; - error4.code = internalError.code; + error3.code = internalError.code; if (internalError.message) { - error4.message = internalError.message; + error3.message = internalError.message; } if (defaultBodyMapper) { - error4.response.parsedBody = deserializedError; + error3.response.parsedBody = deserializedError; } } if (parsedResponse.headers && defaultHeadersMapper) { - error4.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); + error3.response.parsedHeaders = operationSpec.serializer.deserialize(defaultHeadersMapper, parsedResponse.headers.toJSON(), "operationRes.parsedHeaders"); } } catch (defaultError) { - error4.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; + error3.message = `Error "${defaultError.message}" occurred in deserializing the responseBody - "${parsedResponse.bodyAsText}" for the default response.`; } - return { error: error4, shouldReturnResponse: false }; + return { error: error3, shouldReturnResponse: false }; } async function parse(jsonContentTypes, xmlContentTypes, operationResponse, opts, parseXML) { var _a; @@ -41297,8 +41296,8 @@ var require_serializationPolicy = __commonJS({ request.body = JSON.stringify(request.body); } } - } catch (error4) { - throw new Error(`Error "${error4.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); + } catch (error3) { + throw new Error(`Error "${error3.message}" occurred in serializing the payload - ${JSON.stringify(serializedName, void 0, " ")}.`); } } else if (operationSpec.formDataParameters && operationSpec.formDataParameters.length > 0) { request.formData = {}; @@ -41704,16 +41703,16 @@ var require_serviceClient = __commonJS({ options.onResponse(rawResponse, flatResponse); } return flatResponse; - } catch (error4) { - if (typeof error4 === "object" && (error4 === null || error4 === void 0 ? void 0 : error4.response)) { - const rawResponse = error4.response; - const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error4.statusCode] || operationSpec.responses["default"]); - error4.details = flatResponse; + } catch (error3) { + if (typeof error3 === "object" && (error3 === null || error3 === void 0 ? void 0 : error3.response)) { + const rawResponse = error3.response; + const flatResponse = (0, utils_js_1.flattenResponse)(rawResponse, operationSpec.responses[error3.statusCode] || operationSpec.responses["default"]); + error3.details = flatResponse; if (options === null || options === void 0 ? void 0 : options.onResponse) { - options.onResponse(rawResponse, flatResponse, error4); + options.onResponse(rawResponse, flatResponse, error3); } } - throw error4; + throw error3; } } }; @@ -42257,10 +42256,10 @@ var require_extendedClient = __commonJS({ var _a; const userProvidedCallBack = (_a = operationArguments === null || operationArguments === void 0 ? void 0 : operationArguments.options) === null || _a === void 0 ? void 0 : _a.onResponse; let lastResponse; - function onResponse(rawResponse, flatResponse, error4) { + function onResponse(rawResponse, flatResponse, error3) { lastResponse = rawResponse; if (userProvidedCallBack) { - userProvidedCallBack(rawResponse, flatResponse, error4); + userProvidedCallBack(rawResponse, flatResponse, error3); } } operationArguments.options = Object.assign(Object.assign({}, operationArguments.options), { onResponse }); @@ -44339,12 +44338,12 @@ var require_dist6 = __commonJS({ } function setStateError(inputs) { const { state, stateProxy, isOperationError: isOperationError2 } = inputs; - return (error4) => { - if (isOperationError2(error4)) { - stateProxy.setError(state, error4); + return (error3) => { + if (isOperationError2(error3)) { + stateProxy.setError(state, error3); stateProxy.setFailed(state); } - throw error4; + throw error3; }; } function appendReadableErrorMessage(currentMessage, innerMessage) { @@ -44609,16 +44608,16 @@ var require_dist6 = __commonJS({ return void 0; } function getErrorFromResponse(response) { - const error4 = response.flatResponse.error; - if (!error4) { + const error3 = response.flatResponse.error; + if (!error3) { logger.warning(`The long-running operation failed but there is no error property in the response's body`); return; } - if (!error4.code || !error4.message) { + if (!error3.code || !error3.message) { logger.warning(`The long-running operation failed but the error property in the response's body doesn't contain code or message`); return; } - return error4; + return error3; } function calculatePollingIntervalFromDate(retryAfterDate) { const timeNow = Math.floor((/* @__PURE__ */ new Date()).getTime()); @@ -44743,7 +44742,7 @@ var require_dist6 = __commonJS({ */ initState: (config) => ({ status: "running", config }), setCanceled: (state) => state.status = "canceled", - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.status = "running", setSucceeded: (state) => state.status = "succeeded", @@ -44909,7 +44908,7 @@ var require_dist6 = __commonJS({ var createStateProxy = () => ({ initState: (config) => ({ config, isStarted: true }), setCanceled: (state) => state.isCancelled = true, - setError: (state, error4) => state.error = error4, + setError: (state, error3) => state.error = error3, setResult: (state, result) => state.result = result, setRunning: (state) => state.isStarted = true, setSucceeded: (state) => state.isCompleted = true, @@ -45150,9 +45149,9 @@ var require_dist6 = __commonJS({ if (this.operation.state.isCancelled) { this.stopped = true; if (!this.resolveOnUnsuccessful) { - const error4 = new PollerCancelledError("Operation was canceled"); - this.reject(error4); - throw error4; + const error3 = new PollerCancelledError("Operation was canceled"); + this.reject(error3); + throw error3; } } if (this.isDone() && this.resolve) { @@ -45860,7 +45859,7 @@ var require_dist7 = __commonJS({ accountName = ""; } return accountName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract accountName with provided information."); } } @@ -46923,26 +46922,26 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; const maxRetryDelayInMs = (_d = options.maxRetryDelayInMs) !== null && _d !== void 0 ? _d : DEFAULT_RETRY_OPTIONS.maxRetryDelayInMs; const secondaryHost = (_e = options.secondaryHost) !== null && _e !== void 0 ? _e : DEFAULT_RETRY_OPTIONS.secondaryHost; const tryTimeoutInMs = (_f = options.tryTimeoutInMs) !== null && _f !== void 0 ? _f : DEFAULT_RETRY_OPTIONS.tryTimeoutInMs; - function shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }) { + function shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }) { var _a2, _b2; if (attempt >= maxTries) { logger.info(`RetryPolicy: Attempt(s) ${attempt} >= maxTries ${maxTries}, no further try.`); return false; } - if (error4) { + if (error3) { for (const retriableError of retriableErrors) { - if (error4.name.toUpperCase().includes(retriableError) || error4.message.toUpperCase().includes(retriableError) || error4.code && error4.code.toString().toUpperCase() === retriableError) { + if (error3.name.toUpperCase().includes(retriableError) || error3.message.toUpperCase().includes(retriableError) || error3.code && error3.code.toString().toUpperCase() === retriableError) { logger.info(`RetryPolicy: Network error ${retriableError} found, will retry.`); return true; } } - if ((error4 === null || error4 === void 0 ? void 0 : error4.code) === "PARSE_ERROR" && (error4 === null || error4 === void 0 ? void 0 : error4.message.startsWith(`Error "Error: Unclosed root tag`))) { + if ((error3 === null || error3 === void 0 ? void 0 : error3.code) === "PARSE_ERROR" && (error3 === null || error3 === void 0 ? void 0 : error3.message.startsWith(`Error "Error: Unclosed root tag`))) { logger.info("RetryPolicy: Incomplete XML response likely due to service timeout, will retry."); return true; } } - if (response || error4) { - const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error4 === null || error4 === void 0 ? void 0 : error4.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; + if (response || error3) { + const statusCode = (_b2 = (_a2 = response === null || response === void 0 ? void 0 : response.status) !== null && _a2 !== void 0 ? _a2 : error3 === null || error3 === void 0 ? void 0 : error3.statusCode) !== null && _b2 !== void 0 ? _b2 : 0; if (!isPrimaryRetry && statusCode === 404) { logger.info(`RetryPolicy: Secondary access with 404, will retry.`); return true; @@ -46983,12 +46982,12 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; let attempt = 1; let retryAgain = true; let response; - let error4; + let error3; while (retryAgain) { const isPrimaryRetry = secondaryHas404 || !secondaryUrl || !["GET", "HEAD", "OPTIONS"].includes(request.method) || attempt % 2 === 1; request.url = isPrimaryRetry ? primaryUrl : secondaryUrl; response = void 0; - error4 = void 0; + error3 = void 0; try { logger.info(`RetryPolicy: =====> Try=${attempt} ${isPrimaryRetry ? "Primary" : "Secondary"}`); response = await next(request); @@ -46996,13 +46995,13 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; } catch (e) { if (coreRestPipeline.isRestError(e)) { logger.error(`RetryPolicy: Caught error, message: ${e.message}, code: ${e.code}`); - error4 = e; + error3 = e; } else { logger.error(`RetryPolicy: Caught error, message: ${coreUtil.getErrorMessage(e)}`); throw e; } } - retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error4 }); + retryAgain = shouldRetry({ isPrimaryRetry, attempt, response, error: error3 }); if (retryAgain) { await delay2(calculateDelay(isPrimaryRetry, attempt), request.abortSignal, RETRY_ABORT_ERROR); } @@ -47011,7 +47010,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (response) { return response; } - throw error4 !== null && error4 !== void 0 ? error4 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); + throw error3 !== null && error3 !== void 0 ? error3 : new coreRestPipeline.RestError("RetryPolicy failed without known error."); } }; } @@ -61617,8 +61616,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source = newSource; this.setSourceEventHandlers(); return; - }).catch((error4) => { - this.destroy(error4); + }).catch((error3) => { + this.destroy(error3); }); } else { this.destroy(new Error(`Data corruption failure: received less data than required and reached maxRetires limitation. Received data offset: ${this.offset - 1}, data needed offset: ${this.end}, retries: ${this.retries}, max retries: ${this.maxRetryRequests}`)); @@ -61652,10 +61651,10 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.source.removeListener("error", this.sourceErrorOrEndHandler); this.source.removeListener("aborted", this.sourceAbortedHandler); } - _destroy(error4, callback) { + _destroy(error3, callback) { this.removeSourceEventHandlers(); this.source.destroy(); - callback(error4 === null ? void 0 : error4); + callback(error3 === null ? void 0 : error3); } }; var BlobDownloadResponse = class { @@ -63239,8 +63238,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.actives--; this.completed++; this.parallelExecute(); - } catch (error4) { - this.emitter.emit("error", error4); + } catch (error3) { + this.emitter.emit("error", error3); } }); } @@ -63255,9 +63254,9 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; this.parallelExecute(); return new Promise((resolve6, reject) => { this.emitter.on("finish", resolve6); - this.emitter.on("error", (error4) => { + this.emitter.on("error", (error3) => { this.state = BatchStates.Error; - reject(error4); + reject(error3); }); }); } @@ -64358,8 +64357,8 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; if (!buffer2) { try { buffer2 = Buffer.alloc(count); - } catch (error4) { - throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error4.message}`); + } catch (error3) { + throw new Error(`Unable to allocate the buffer of size: ${count}(in bytes). Please try passing your own buffer to the "downloadToBuffer" method or try using other methods like "download" or "downloadToFile". ${error3.message}`); } } if (buffer2.length < count) { @@ -64443,7 +64442,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return { blobName, containerName }; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract blobName and containerName with provided information."); } } @@ -67595,7 +67594,7 @@ ${key}:${decodeURIComponent(lowercaseQueries[key])}`; throw new Error("Provided containerName is invalid."); } return containerName; - } catch (error4) { + } catch (error3) { throw new Error("Unable to extract containerName with provided information."); } } @@ -68962,9 +68961,9 @@ var require_uploadUtils = __commonJS({ throw new errors_1.InvalidResponseError(`uploadCacheArchiveSDK: upload failed with status code ${response._response.status}`); } return response; - } catch (error4) { - core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error4.message}`); - throw error4; + } catch (error3) { + core14.warning(`uploadCacheArchiveSDK: internal error uploading cache archive: ${error3.message}`); + throw error3; } finally { uploadProgress.stopDisplayTimer(); } @@ -69078,12 +69077,12 @@ var require_requestUtils = __commonJS({ let isRetryable = false; try { response = yield method(); - } catch (error4) { + } catch (error3) { if (onError) { - response = onError(error4); + response = onError(error3); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (response) { statusCode = getStatusCode(response); @@ -69117,13 +69116,13 @@ var require_requestUtils = __commonJS({ delay2, // If the error object contains the statusCode property, extract it and return // an TypedResponse so it can be processed by the retry logic. - (error4) => { - if (error4 instanceof http_client_1.HttpClientError) { + (error3) => { + if (error3 instanceof http_client_1.HttpClientError) { return { - statusCode: error4.statusCode, + statusCode: error3.statusCode, result: null, headers: {}, - error: error4 + error: error3 }; } else { return void 0; @@ -69939,8 +69938,8 @@ Other caches with similar key:`); start, end, autoClose: false - }).on("error", (error4) => { - throw new Error(`Cache upload failed because file read failed with ${error4.message}`); + }).on("error", (error3) => { + throw new Error(`Cache upload failed because file read failed with ${error3.message}`); }), start, end); } }))); @@ -71786,8 +71785,8 @@ var require_reflection_json_reader = __commonJS({ break; return base64_1.base64decode(json2); } - } catch (error4) { - e = error4.message; + } catch (error3) { + e = error3.message; } this.assert(false, fieldName + (e ? " - " + e : ""), json2); } @@ -73358,12 +73357,12 @@ var require_rpc_output_stream = __commonJS({ * at a time. * Can be used to wrap a stream by using the other stream's `onNext`. */ - notifyNext(message, error4, complete) { - runtime_1.assert((message ? 1 : 0) + (error4 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); + notifyNext(message, error3, complete) { + runtime_1.assert((message ? 1 : 0) + (error3 ? 1 : 0) + (complete ? 1 : 0) <= 1, "only one emission at a time"); if (message) this.notifyMessage(message); - if (error4) - this.notifyError(error4); + if (error3) + this.notifyError(error3); if (complete) this.notifyComplete(); } @@ -73383,12 +73382,12 @@ var require_rpc_output_stream = __commonJS({ * * Triggers onNext and onError callbacks. */ - notifyError(error4) { + notifyError(error3) { runtime_1.assert(!this.closed, "stream is closed"); - this._closed = error4; - this.pushIt(error4); - this._lis.err.forEach((l) => l(error4)); - this._lis.nxt.forEach((l) => l(void 0, error4, false)); + this._closed = error3; + this.pushIt(error3); + this._lis.err.forEach((l) => l(error3)); + this._lis.nxt.forEach((l) => l(void 0, error3, false)); this.clearLis(); } /** @@ -73852,8 +73851,8 @@ var require_test_transport = __commonJS({ } try { yield delay2(this.responseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } if (this.data.response instanceof rpc_error_1.RpcError) { @@ -73864,8 +73863,8 @@ var require_test_transport = __commonJS({ stream2.notifyMessage(msg); try { yield delay2(this.betweenResponseDelay, abort)(void 0); - } catch (error4) { - stream2.notifyError(error4); + } catch (error3) { + stream2.notifyError(error3); return; } } @@ -74928,8 +74927,8 @@ var require_util10 = __commonJS({ (0, core_1.setSecret)(signature); (0, core_1.setSecret)(encodeURIComponent(signature)); } - } catch (error4) { - (0, core_1.debug)(`Failed to parse URL: ${url2} ${error4 instanceof Error ? error4.message : String(error4)}`); + } catch (error3) { + (0, core_1.debug)(`Failed to parse URL: ${url2} ${error3 instanceof Error ? error3.message : String(error3)}`); } } exports2.maskSigUrl = maskSigUrl; @@ -75025,8 +75024,8 @@ var require_cacheTwirpClient = __commonJS({ return this.httpClient.post(url2, JSON.stringify(data), headers); })); return body; - } catch (error4) { - throw new Error(`Failed to ${method}: ${error4.message}`); + } catch (error3) { + throw new Error(`Failed to ${method}: ${error3.message}`); } }); } @@ -75057,18 +75056,18 @@ var require_cacheTwirpClient = __commonJS({ } errorMessage = `${errorMessage}: ${body.msg}`; } - } catch (error4) { - if (error4 instanceof SyntaxError) { + } catch (error3) { + if (error3 instanceof SyntaxError) { (0, core_1.debug)(`Raw Body: ${rawBody}`); } - if (error4 instanceof errors_1.UsageError) { - throw error4; + if (error3 instanceof errors_1.UsageError) { + throw error3; } - if (errors_1.NetworkError.isNetworkErrorCode(error4 === null || error4 === void 0 ? void 0 : error4.code)) { - throw new errors_1.NetworkError(error4 === null || error4 === void 0 ? void 0 : error4.code); + if (errors_1.NetworkError.isNetworkErrorCode(error3 === null || error3 === void 0 ? void 0 : error3.code)) { + throw new errors_1.NetworkError(error3 === null || error3 === void 0 ? void 0 : error3.code); } isRetryable = true; - errorMessage = error4.message; + errorMessage = error3.message; } if (!isRetryable) { throw new Error(`Received non-retryable error: ${errorMessage}`); @@ -75336,8 +75335,8 @@ var require_tar = __commonJS({ cwd, env: Object.assign(Object.assign({}, process.env), { MSYS: "winsymlinks:nativestrict" }) }); - } catch (error4) { - throw new Error(`${command.split(" ")[0]} failed with error: ${error4 === null || error4 === void 0 ? void 0 : error4.message}`); + } catch (error3) { + throw new Error(`${command.split(" ")[0]} failed with error: ${error3 === null || error3 === void 0 ? void 0 : error3.message}`); } } }); @@ -75538,22 +75537,22 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return cacheEntry.cacheKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -75608,15 +75607,15 @@ var require_cache3 = __commonJS({ yield (0, tar_1.extractTar)(archivePath, compressionMethod); core14.info("Cache restored successfully"); return response.matchedKey; - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else { if (typedError instanceof http_client_1.HttpClientError && typeof typedError.statusCode === "number" && typedError.statusCode >= 500) { - core14.error(`Failed to restore: ${error4.message}`); + core14.error(`Failed to restore: ${error3.message}`); } else { - core14.warning(`Failed to restore: ${error4.message}`); + core14.warning(`Failed to restore: ${error3.message}`); } } } finally { @@ -75624,8 +75623,8 @@ var require_cache3 = __commonJS({ if (archivePath) { yield utils.unlinkFile(archivePath); } - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return void 0; @@ -75687,10 +75686,10 @@ var require_cache3 = __commonJS({ } core14.debug(`Saving Cache (ID: ${cacheId})`); yield cacheHttpClient.saveCache(cacheId, archivePath, "", options); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core14.info(`Failed to save: ${typedError.message}`); } else { @@ -75703,8 +75702,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -75749,8 +75748,8 @@ var require_cache3 = __commonJS({ throw new Error(response.message || "Response was not ok"); } signedUploadUrl = response.signedUploadUrl; - } catch (error4) { - core14.debug(`Failed to reserve cache: ${error4}`); + } catch (error3) { + core14.debug(`Failed to reserve cache: ${error3}`); throw new ReserveCacheError(`Unable to reserve cache with key ${key}, another job may be creating this cache.`); } core14.debug(`Attempting to upload cache located at: ${archivePath}`); @@ -75769,10 +75768,10 @@ var require_cache3 = __commonJS({ throw new Error(`Unable to finalize cache with key ${key}, another job may be finalizing this cache.`); } cacheId = parseInt(finalizeResponse.entryId); - } catch (error4) { - const typedError = error4; + } catch (error3) { + const typedError = error3; if (typedError.name === ValidationError.name) { - throw error4; + throw error3; } else if (typedError.name === ReserveCacheError.name) { core14.info(`Failed to save: ${typedError.message}`); } else if (typedError.name === FinalizeCacheError.name) { @@ -75787,8 +75786,8 @@ var require_cache3 = __commonJS({ } finally { try { yield utils.unlinkFile(archivePath); - } catch (error4) { - core14.debug(`Failed to delete archive: ${error4}`); + } catch (error3) { + core14.debug(`Failed to delete archive: ${error3}`); } } return cacheId; @@ -79811,7 +79810,7 @@ var require_debug2 = __commonJS({ if (!debug6) { try { debug6 = require_src()("follow-redirects"); - } catch (error4) { + } catch (error3) { } if (typeof debug6 !== "function") { debug6 = function() { @@ -79844,8 +79843,8 @@ var require_follow_redirects = __commonJS({ var useNativeURL = false; try { assert(new URL2("")); - } catch (error4) { - useNativeURL = error4.code === "ERR_INVALID_URL"; + } catch (error3) { + useNativeURL = error3.code === "ERR_INVALID_URL"; } var preservedUrlFields = [ "auth", @@ -79919,9 +79918,9 @@ var require_follow_redirects = __commonJS({ this._currentRequest.abort(); this.emit("abort"); }; - RedirectableRequest.prototype.destroy = function(error4) { - destroyRequest(this._currentRequest, error4); - destroy.call(this, error4); + RedirectableRequest.prototype.destroy = function(error3) { + destroyRequest(this._currentRequest, error3); + destroy.call(this, error3); return this; }; RedirectableRequest.prototype.write = function(data, encoding, callback) { @@ -80088,10 +80087,10 @@ var require_follow_redirects = __commonJS({ var i = 0; var self2 = this; var buffers = this._requestBodyBuffers; - (function writeNext(error4) { + (function writeNext(error3) { if (request === self2._currentRequest) { - if (error4) { - self2.emit("error", error4); + if (error3) { + self2.emit("error", error3); } else if (i < buffers.length) { var buffer = buffers[i++]; if (!request.finished) { @@ -80290,12 +80289,12 @@ var require_follow_redirects = __commonJS({ }); return CustomError; } - function destroyRequest(request, error4) { + function destroyRequest(request, error3) { for (var event of events) { request.removeListener(event, eventHandlers[event]); } request.on("error", noop); - request.destroy(error4); + request.destroy(error3); } function isSubdomain(subdomain, domain) { assert(isString(subdomain) && isString(domain)); @@ -83231,7 +83230,6 @@ var fs = __toESM(require("fs")); var fsPromises = __toESM(require("fs/promises")); var path = __toESM(require("path")); var core3 = __toESM(require_core()); -var exec = __toESM(require_exec()); var io = __toESM(require_io2()); // node_modules/get-folder-size/index.js @@ -83249,14 +83247,14 @@ async function core(rootItemPath, options = {}, returnType = {}) { await processItem(rootItemPath); async function processItem(itemPath) { if (options.ignore?.test(itemPath)) return; - const stats = returnType.strict ? await fs13.lstat(itemPath, { bigint: true }) : await fs13.lstat(itemPath, { bigint: true }).catch((error4) => errors.push(error4)); + const stats = returnType.strict ? await fs13.lstat(itemPath, { bigint: true }) : await fs13.lstat(itemPath, { bigint: true }).catch((error3) => errors.push(error3)); if (typeof stats !== "object") return; if (!foundInos.has(stats.ino)) { foundInos.add(stats.ino); folderSize += stats.size; } if (stats.isDirectory()) { - const directoryItems = returnType.strict ? await fs13.readdir(itemPath) : await fs13.readdir(itemPath).catch((error4) => errors.push(error4)); + const directoryItems = returnType.strict ? await fs13.readdir(itemPath) : await fs13.readdir(itemPath).catch((error3) => errors.push(error3)); if (typeof directoryItems !== "object") return; await Promise.all( directoryItems.map( @@ -83267,13 +83265,13 @@ async function core(rootItemPath, options = {}, returnType = {}) { } if (!options.bigint) { if (folderSize > BigInt(Number.MAX_SAFE_INTEGER)) { - const error4 = new RangeError( + const error3 = new RangeError( "The folder size is too large to return as a Number. You can instruct this package to return a BigInt instead." ); if (returnType.strict) { - throw error4; + throw error3; } - errors.push(error4); + errors.push(error3); folderSize = Number.MAX_SAFE_INTEGER; } else { folderSize = Number(folderSize); @@ -84153,6 +84151,18 @@ function charFromCodepoint(c) { (c - 65536 & 1023) + 56320 ); } +function setProperty(object, key, value) { + if (key === "__proto__") { + Object.defineProperty(object, key, { + configurable: true, + enumerable: true, + writable: true, + value + }); + } else { + object[key] = value; + } +} var simpleEscapeCheck = new Array(256); var simpleEscapeMap = new Array(256); for (i = 0; i < 256; i++) { @@ -84272,7 +84282,7 @@ function mergeMappings(state, destination, source, overridableKeys) { for (index = 0, quantity = sourceKeys.length; index < quantity; index += 1) { key = sourceKeys[index]; if (!_hasOwnProperty$1.call(destination, key)) { - destination[key] = source[key]; + setProperty(destination, key, source[key]); overridableKeys[key] = true; } } @@ -84312,16 +84322,7 @@ function storeMappingPair(state, _result, overridableKeys, keyTag, keyNode, valu state.position = startPos || state.position; throwError(state, "duplicated mapping key"); } - if (keyNode === "__proto__") { - Object.defineProperty(_result, keyNode, { - configurable: true, - enumerable: true, - writable: true, - value: valueNode - }); - } else { - _result[keyNode] = valueNode; - } + setProperty(_result, keyNode, valueNode); delete overridableKeys[keyNode]; } return _result; @@ -85886,9 +85887,9 @@ function getExtraOptionsEnvParam() { try { return load(raw); } catch (unwrappedError) { - const error4 = wrapError(unwrappedError); + const error3 = wrapError(unwrappedError); throw new ConfigurationError( - `${varName} environment variable is set, but does not contain valid JSON: ${error4.message}` + `${varName} environment variable is set, but does not contain valid JSON: ${error3.message}` ); } } @@ -86031,11 +86032,11 @@ function parseMatrixInput(matrixInput) { } return JSON.parse(matrixInput); } -function wrapError(error4) { - return error4 instanceof Error ? error4 : new Error(String(error4)); +function wrapError(error3) { + return error3 instanceof Error ? error3 : new Error(String(error3)); } -function getErrorMessage(error4) { - return error4 instanceof Error ? error4.message : String(error4); +function getErrorMessage(error3) { + return error3 instanceof Error ? error3.message : String(error3); } async function checkDiskUsage(logger) { try { @@ -86058,9 +86059,9 @@ async function checkDiskUsage(logger) { numAvailableBytes: diskUsage.bavail * blockSizeInBytes, numTotalBytes: diskUsage.blocks * blockSizeInBytes }; - } catch (error4) { + } catch (error3) { logger.warning( - `Failed to check available disk space: ${getErrorMessage(error4)}` + `Failed to check available disk space: ${getErrorMessage(error3)}` ); return void 0; } @@ -86072,7 +86073,7 @@ function checkActionVersion(version, githubVersion) { semver.coerce(githubVersion.version) ?? "0.0.0", ">=3.20" )) { - core3.error( + core3.warning( "CodeQL Action v3 will be deprecated in December 2026. Please update all occurrences of the CodeQL Action in your workflow files to v4. For more information, see https://github.blog/changelog/2025-10-28-upcoming-deprecation-of-codeql-action-v3/" ); core3.exportVariable("CODEQL_ACTION_DID_LOG_VERSION_DEPRECATION" /* LOG_VERSION_DEPRECATION */, "true"); @@ -86504,8 +86505,8 @@ var path4 = __toESM(require("path")); var semver3 = __toESM(require_semver2()); // src/defaults.json -var bundleVersion = "codeql-bundle-v2.23.3"; -var cliVersion = "2.23.3"; +var bundleVersion = "codeql-bundle-v2.23.6"; +var cliVersion = "2.23.6"; // src/overlay-database-utils.ts var fs3 = __toESM(require("fs")); @@ -86537,13 +86538,13 @@ var runGitCommand = async function(workingDirectory, args, customErrorMessage) { cwd: workingDirectory }).exec(); return stdout; - } catch (error4) { + } catch (error3) { let reason = stderr; if (stderr.includes("not a git repository")) { reason = "The checkout path provided to the action does not appear to be a git repository."; } core7.info(`git call failed. ${customErrorMessage} Error: ${reason}`); - throw error4; + throw error3; } }; var getCommitOid = async function(checkoutPath, ref = "HEAD") { @@ -86738,7 +86739,7 @@ function formatDuration(durationMs) { } // src/overlay-database-utils.ts -var CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +var CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB = 7500; var OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_BYTES = OVERLAY_BASE_DATABASE_MAX_UPLOAD_SIZE_MB * 1e6; async function writeBaseDatabaseOidsFile(config, sourceRoot) { @@ -86868,6 +86869,11 @@ var featureConfig = { legacyApi: true, minimumVersion: void 0 }, + ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0" + }, ["overlay_analysis" /* OverlayAnalysis */]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -86979,21 +86985,21 @@ var featureConfig = { minimumVersion: void 0, toolsFeature: "pythonDefaultIsToNotExtractStdlib" /* PythonDefaultIsToNotExtractStdlib */ }, - ["use_repository_properties" /* UseRepositoryProperties */]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: void 0 - }, ["qa_telemetry_enabled" /* QaTelemetryEnabled */]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: void 0 }, - ["java_minimize_dependency_jars" /* JavaMinimizeDependencyJars */]: { + ["upload_overlay_db_to_api" /* UploadOverlayDbToApi */]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0" + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: void 0 + }, + ["use_repository_properties" /* UseRepositoryProperties */]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: void 0 }, ["validate_db_config" /* ValidateDbConfig */]: { defaultValue: false, @@ -87309,6 +87315,8 @@ ${jsonContents}` var actionsCache2 = __toESM(require_cache3()); // src/config-utils.ts +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 2e4; +var OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1e6; var OVERLAY_ANALYSIS_FEATURES = { actions: "overlay_analysis_actions" /* OverlayAnalysisActions */, cpp: "overlay_analysis_cpp" /* OverlayAnalysisCpp */, @@ -87386,9 +87394,9 @@ function isFirstPartyAnalysis(actionName) { function isThirdPartyAnalysis(actionName) { return !isFirstPartyAnalysis(actionName); } -function getActionsStatus(error4, otherFailureCause) { - if (error4 || otherFailureCause) { - return error4 instanceof ConfigurationError ? "user-error" : "failure"; +function getActionsStatus(error3, otherFailureCause) { + if (error3 || otherFailureCause) { + return error3 instanceof ConfigurationError ? "user-error" : "failure"; } else { return "success"; } @@ -87602,19 +87610,19 @@ var CliError = class extends Error { this.stderr = stderr; } }; -function extractFatalErrors(error4) { +function extractFatalErrors(error3) { const fatalErrorRegex = /.*fatal (internal )?error occurr?ed(. Details)?:/gi; let fatalErrors = []; let lastFatalErrorIndex; let match; - while ((match = fatalErrorRegex.exec(error4)) !== null) { + while ((match = fatalErrorRegex.exec(error3)) !== null) { if (lastFatalErrorIndex !== void 0) { - fatalErrors.push(error4.slice(lastFatalErrorIndex, match.index).trim()); + fatalErrors.push(error3.slice(lastFatalErrorIndex, match.index).trim()); } lastFatalErrorIndex = match.index; } if (lastFatalErrorIndex !== void 0) { - const lastError = error4.slice(lastFatalErrorIndex).trim(); + const lastError = error3.slice(lastFatalErrorIndex).trim(); if (fatalErrors.length === 0) { return lastError; } @@ -87630,9 +87638,9 @@ function extractFatalErrors(error4) { } return void 0; } -function extractAutobuildErrors(error4) { +function extractAutobuildErrors(error3) { const pattern = /.*\[autobuild\] \[ERROR\] (.*)/gi; - let errorLines = [...error4.matchAll(pattern)].map((match) => match[1]); + let errorLines = [...error3.matchAll(pattern)].map((match) => match[1]); if (errorLines.length > 10) { errorLines = errorLines.slice(0, 10); errorLines.push("(truncated)"); @@ -90808,15 +90816,15 @@ function validateSarifFileSchema(sarif, sarifFilePath, logger) { const warnings = (result.errors ?? []).filter( (err) => err.name === "format" && typeof err.argument === "string" && warningAttributes.includes(err.argument) ); - for (const warning9 of warnings) { + for (const warning10 of warnings) { logger.info( - `Warning: '${warning9.instance}' is not a valid URI in '${warning9.property}'.` + `Warning: '${warning10.instance}' is not a valid URI in '${warning10.property}'.` ); } if (errors.length > 0) { - for (const error4 of errors) { - logger.startGroup(`Error details: ${error4.stack}`); - logger.info(JSON.stringify(error4, null, 2)); + for (const error3 of errors) { + logger.startGroup(`Error details: ${error3.stack}`); + logger.info(JSON.stringify(error3, null, 2)); logger.endGroup(); } const sarifErrors = errors.map((e) => `- ${e.stack}`); @@ -91039,10 +91047,10 @@ function shouldConsiderConfigurationError(processingErrors) { } function shouldConsiderInvalidRequest(processingErrors) { return processingErrors.every( - (error4) => error4.startsWith("rejecting SARIF") || error4.startsWith("an invalid URI was provided as a SARIF location") || error4.startsWith("locationFromSarifResult: expected artifact location") || error4.startsWith( + (error3) => error3.startsWith("rejecting SARIF") || error3.startsWith("an invalid URI was provided as a SARIF location") || error3.startsWith("locationFromSarifResult: expected artifact location") || error3.startsWith( "could not convert rules: invalid security severity value, is not a number" ) || /^SARIF URI scheme [^\s]* did not match the checkout URI scheme [^\s]*/.test( - error4 + error3 ) ); } @@ -91241,18 +91249,18 @@ async function run() { logger ); } catch (unwrappedError) { - const error4 = isThirdPartyAnalysis("upload-sarif" /* UploadSarif */) && unwrappedError instanceof InvalidSarifUploadError ? new ConfigurationError(unwrappedError.message) : wrapError(unwrappedError); - const message = error4.message; + const error3 = isThirdPartyAnalysis("upload-sarif" /* UploadSarif */) && unwrappedError instanceof InvalidSarifUploadError ? new ConfigurationError(unwrappedError.message) : wrapError(unwrappedError); + const message = error3.message; core13.setFailed(message); const errorStatusReportBase = await createStatusReportBase( "upload-sarif" /* UploadSarif */, - getActionsStatus(error4), + getActionsStatus(error3), startedAt, void 0, await checkDiskUsage(logger), logger, message, - error4.stack + error3.stack ); if (errorStatusReportBase !== void 0) { await sendStatusReport(errorStatusReportBase); @@ -91263,9 +91271,9 @@ async function run() { async function runWrapper() { try { await run(); - } catch (error4) { + } catch (error3) { core13.setFailed( - `codeql/upload-sarif action failed: ${getErrorMessage(error4)}` + `codeql/upload-sarif action failed: ${getErrorMessage(error3)}` ); } } @@ -91279,7 +91287,7 @@ undici/lib/websocket/frame.js: (*! ws. MIT License. Einar Otto Stangvik *) js-yaml/dist/js-yaml.mjs: - (*! js-yaml 4.1.0 https://github.com/nodeca/js-yaml @license MIT *) + (*! js-yaml 4.1.1 https://github.com/nodeca/js-yaml @license MIT *) long/index.js: (** diff --git a/package-lock.json b/package-lock.json index 78a1b05a6..4c6ca8624 100644 --- a/package-lock.json +++ b/package-lock.json @@ -1,12 +1,12 @@ { "name": "codeql", - "version": "4.31.3", + "version": "4.31.6", "lockfileVersion": 3, "requires": true, "packages": { "": { "name": "codeql", - "version": "4.31.3", + "version": "4.31.6", "license": "MIT", "dependencies": { "@actions/artifact": "^4.0.0", @@ -20,23 +20,21 @@ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", "archiver": "^7.0.1", "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", "jsonschema": "1.4.1", "long": "^5.3.2", "node-forge": "^1.3.1", - "octokit": "^5.0.5", "semver": "^7.7.3", "uuid": "^13.0.0" }, "devDependencies": { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -44,10 +42,10 @@ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", "ava": "^6.4.1", @@ -57,9 +55,9 @@ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", - "glob": "^11.0.3", + "glob": "^11.1.0", "nock": "^14.0.10", "sinon": "^21.0.0", "typescript": "^5.9.3" @@ -1417,16 +1415,16 @@ } }, "node_modules/@eslint/compat": { - "version": "1.4.1", - "resolved": "https://registry.npmjs.org/@eslint/compat/-/compat-1.4.1.tgz", - "integrity": "sha512-cfO82V9zxxGBxcQDr1lfaYB7wykTa0b00mGa36FrJl7iTFd0Z2cHfEYuxcBRP/iNijCsWsEkA+jzT8hGYmv33w==", + "version": "2.0.0", + "resolved": "https://registry.npmjs.org/@eslint/compat/-/compat-2.0.0.tgz", + "integrity": "sha512-T9AfE1G1uv4wwq94ozgTGio5EUQBqAVe1X9qsQtSNVEYW6j3hvtZVm8Smr4qL1qDPFg+lOB2cL5RxTRMzq4CTA==", "dev": true, "license": "Apache-2.0", "dependencies": { - "@eslint/core": "^0.17.0" + "@eslint/core": "^1.0.0" }, "engines": { - "node": "^18.18.0 || ^20.9.0 || >=21.1.0" + "node": "^20.19.0 || ^22.13.0 || >=24" }, "peerDependencies": { "eslint": "^8.40 || 9" @@ -1438,16 +1436,16 @@ } }, "node_modules/@eslint/core": { - "version": "0.17.0", - "resolved": "https://registry.npmjs.org/@eslint/core/-/core-0.17.0.tgz", - "integrity": "sha512-yL/sLrpmtDaFEiUj1osRP4TI2MDz1AddJL+jZ7KSqvBuliN4xqYY54IfdN8qD8Toa6g1iloph1fxQNkjOxrrpQ==", + "version": "1.0.0", + "resolved": "https://registry.npmjs.org/@eslint/core/-/core-1.0.0.tgz", + "integrity": "sha512-PRfWP+8FOldvbApr6xL7mNCw4cJcSTq4GA7tYbgq15mRb0kWKO/wEB2jr+uwjFH3sZvEZneZyCUGTxsv4Sahyw==", "dev": true, "license": "Apache-2.0", "dependencies": { "@types/json-schema": "^7.0.15" }, "engines": { - "node": "^18.18.0 || ^20.9.0 || >=21.1.0" + "node": "^20.19.0 || ^22.13.0 || >=24" } }, "node_modules/@eslint/eslintrc": { @@ -1578,7 +1576,6 @@ "version": "4.0.1", "resolved": "https://registry.npmjs.org/@isaacs/balanced-match/-/balanced-match-4.0.1.tgz", "integrity": "sha512-yzMTt9lEb8Gv7zRioUilSglI0c0smZ9k5D65677DLWLtWJaXIS3CqcGyUFByYKlnUj6TkjLVs54fBl6+TiGQDQ==", - "dev": true, "license": "MIT", "engines": { "node": "20 || >=22" @@ -1588,7 +1585,6 @@ "version": "5.0.0", "resolved": "https://registry.npmjs.org/@isaacs/brace-expansion/-/brace-expansion-5.0.0.tgz", "integrity": "sha512-ZT55BDLV0yv0RBm2czMiZ+SqCGO7AvmOM3G/w2xhVPH+te0aKgFjmBvGlL1dH+ql2tgGO3MVrbb3jCKyvpgnxA==", - "dev": true, "license": "MIT", "dependencies": { "@isaacs/balanced-match": "^4.0.1" @@ -1719,6 +1715,7 @@ "resolved": "https://registry.npmjs.org/@microsoft/eslint-formatter-sarif/-/eslint-formatter-sarif-3.1.0.tgz", "integrity": "sha512-/mn4UXziHzGXnKCg+r8HGgPy+w4RzpgdoqFuqaKOqUVBT5x2CygGefIrO4SusaY7t0C4gyIWMNu6YQT6Jw64Cw==", "dev": true, + "license": "MIT", "dependencies": { "eslint": "^8.9.0", "jschardet": "latest", @@ -1788,182 +1785,6 @@ "node": ">=12.4.0" } }, - "node_modules/@octokit/app": { - "version": "16.1.2", - "resolved": "https://registry.npmjs.org/@octokit/app/-/app-16.1.2.tgz", - "integrity": "sha512-8j7sEpUYVj18dxvh0KWj6W/l6uAiVRBl1JBDVRqH1VHKAO/G5eRVl4yEoYACjakWers1DjUkcCHyJNQK47JqyQ==", - "license": "MIT", - "dependencies": { - "@octokit/auth-app": "^8.1.2", - "@octokit/auth-unauthenticated": "^7.0.3", - "@octokit/core": "^7.0.6", - "@octokit/oauth-app": "^8.0.3", - "@octokit/plugin-paginate-rest": "^14.0.0", - "@octokit/types": "^16.0.0", - "@octokit/webhooks": "^14.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/app/node_modules/@octokit/auth-token": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/@octokit/auth-token/-/auth-token-6.0.0.tgz", - "integrity": "sha512-P4YJBPdPSpWTQ1NU4XYdvHvXJJDxM6YwpS0FZHRgP7YFkdVxsWcpWGy/NVqlAA7PcPCnMacXlRm1y2PFZRWL/w==", - "license": "MIT", - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/app/node_modules/@octokit/core": { - "version": "7.0.6", - "resolved": "https://registry.npmjs.org/@octokit/core/-/core-7.0.6.tgz", - "integrity": "sha512-DhGl4xMVFGVIyMwswXeyzdL4uXD5OGILGX5N8Y+f6W7LhC1Ze2poSNrkF/fedpVDHEEZ+PHFW0vL14I+mm8K3Q==", - "license": "MIT", - "dependencies": { - "@octokit/auth-token": "^6.0.0", - "@octokit/graphql": "^9.0.3", - "@octokit/request": "^10.0.6", - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0", - "before-after-hook": "^4.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/app/node_modules/@octokit/graphql": { - "version": "9.0.3", - "resolved": "https://registry.npmjs.org/@octokit/graphql/-/graphql-9.0.3.tgz", - "integrity": "sha512-grAEuupr/C1rALFnXTv6ZQhFuL1D8G5y8CN04RgrO4FIPMrtm+mcZzFG7dcBm+nq+1ppNixu+Jd78aeJOYxlGA==", - "license": "MIT", - "dependencies": { - "@octokit/request": "^10.0.6", - "@octokit/types": "^16.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/app/node_modules/@octokit/plugin-paginate-rest": { - "version": "14.0.0", - "resolved": "https://registry.npmjs.org/@octokit/plugin-paginate-rest/-/plugin-paginate-rest-14.0.0.tgz", - "integrity": "sha512-fNVRE7ufJiAA3XUrha2omTA39M6IXIc6GIZLvlbsm8QOQCYvpq/LkMNGyFlB1d8hTDzsAXa3OKtybdMAYsV/fw==", - "license": "MIT", - "dependencies": { - "@octokit/types": "^16.0.0" - }, - "engines": { - "node": ">= 20" - }, - "peerDependencies": { - "@octokit/core": ">=6" - } - }, - "node_modules/@octokit/app/node_modules/before-after-hook": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/before-after-hook/-/before-after-hook-4.0.0.tgz", - "integrity": "sha512-q6tR3RPqIB1pMiTRMFcZwuG5T8vwp+vUvEG0vuI6B+Rikh5BfPp2fQ82c925FOs+b0lcFQ8CFrL+KbilfZFhOQ==", - "license": "Apache-2.0" - }, - "node_modules/@octokit/app/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, - "node_modules/@octokit/auth-app": { - "version": "8.1.2", - "resolved": "https://registry.npmjs.org/@octokit/auth-app/-/auth-app-8.1.2.tgz", - "integrity": "sha512-db8VO0PqXxfzI6GdjtgEFHY9tzqUql5xMFXYA12juq8TeTgPAuiiP3zid4h50lwlIP457p5+56PnJOgd2GGBuw==", - "license": "MIT", - "dependencies": { - "@octokit/auth-oauth-app": "^9.0.3", - "@octokit/auth-oauth-user": "^6.0.2", - "@octokit/request": "^10.0.6", - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0", - "toad-cache": "^3.7.0", - "universal-github-app-jwt": "^2.2.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/auth-app/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, - "node_modules/@octokit/auth-oauth-app": { - "version": "9.0.3", - "resolved": "https://registry.npmjs.org/@octokit/auth-oauth-app/-/auth-oauth-app-9.0.3.tgz", - "integrity": "sha512-+yoFQquaF8OxJSxTb7rnytBIC2ZLbLqA/yb71I4ZXT9+Slw4TziV9j/kyGhUFRRTF2+7WlnIWsePZCWHs+OGjg==", - "license": "MIT", - "dependencies": { - "@octokit/auth-oauth-device": "^8.0.3", - "@octokit/auth-oauth-user": "^6.0.2", - "@octokit/request": "^10.0.6", - "@octokit/types": "^16.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/auth-oauth-app/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, - "node_modules/@octokit/auth-oauth-device": { - "version": "8.0.3", - "resolved": "https://registry.npmjs.org/@octokit/auth-oauth-device/-/auth-oauth-device-8.0.3.tgz", - "integrity": "sha512-zh2W0mKKMh/VWZhSqlaCzY7qFyrgd9oTWmTmHaXnHNeQRCZr/CXy2jCgHo4e4dJVTiuxP5dLa0YM5p5QVhJHbw==", - "license": "MIT", - "dependencies": { - "@octokit/oauth-methods": "^6.0.2", - "@octokit/request": "^10.0.6", - "@octokit/types": "^16.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/auth-oauth-device/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, - "node_modules/@octokit/auth-oauth-user": { - "version": "6.0.2", - "resolved": "https://registry.npmjs.org/@octokit/auth-oauth-user/-/auth-oauth-user-6.0.2.tgz", - "integrity": "sha512-qLoPPc6E6GJoz3XeDG/pnDhJpTkODTGG4kY0/Py154i/I003O9NazkrwJwRuzgCalhzyIeWQ+6MDvkUmKXjg/A==", - "license": "MIT", - "dependencies": { - "@octokit/auth-oauth-device": "^8.0.3", - "@octokit/oauth-methods": "^6.0.2", - "@octokit/request": "^10.0.6", - "@octokit/types": "^16.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/auth-oauth-user/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, "node_modules/@octokit/auth-token": { "version": "4.0.0", "resolved": "https://registry.npmjs.org/@octokit/auth-token/-/auth-token-4.0.0.tgz", @@ -1972,19 +1793,6 @@ "node": ">= 18" } }, - "node_modules/@octokit/auth-unauthenticated": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/@octokit/auth-unauthenticated/-/auth-unauthenticated-7.0.3.tgz", - "integrity": "sha512-8Jb1mtUdmBHL7lGmop9mU9ArMRUTRhg8vp0T1VtZ4yd9vEm3zcLwmjQkhNEduKawOOORie61xhtYIhTDN+ZQ3g==", - "license": "MIT", - "dependencies": { - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0" - }, - "engines": { - "node": ">= 20" - } - }, "node_modules/@octokit/core": { "version": "5.2.2", "resolved": "https://registry.npmjs.org/@octokit/core/-/core-5.2.2.tgz", @@ -2055,25 +1863,6 @@ "@octokit/openapi-types": "^24.2.0" } }, - "node_modules/@octokit/endpoint": { - "version": "11.0.2", - "resolved": "https://registry.npmjs.org/@octokit/endpoint/-/endpoint-11.0.2.tgz", - "integrity": "sha512-4zCpzP1fWc7QlqunZ5bSEjxc6yLAlRTnDwKtgXfcI/FxxGoqedDG8V2+xJ60bV2kODqcGB+nATdtap/XYq2NZQ==", - "license": "MIT", - "dependencies": { - "@octokit/types": "^16.0.0", - "universal-user-agent": "^7.0.2" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/endpoint/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, "node_modules/@octokit/graphql": { "version": "7.1.1", "resolved": "https://registry.npmjs.org/@octokit/graphql/-/graphql-7.1.1.tgz", @@ -2139,112 +1928,11 @@ "@octokit/openapi-types": "^24.2.0" } }, - "node_modules/@octokit/oauth-app": { - "version": "8.0.3", - "resolved": "https://registry.npmjs.org/@octokit/oauth-app/-/oauth-app-8.0.3.tgz", - "integrity": "sha512-jnAjvTsPepyUaMu9e69hYBuozEPgYqP4Z3UnpmvoIzHDpf8EXDGvTY1l1jK0RsZ194oRd+k6Hm13oRU8EoDFwg==", - "license": "MIT", - "dependencies": { - "@octokit/auth-oauth-app": "^9.0.2", - "@octokit/auth-oauth-user": "^6.0.1", - "@octokit/auth-unauthenticated": "^7.0.2", - "@octokit/core": "^7.0.5", - "@octokit/oauth-authorization-url": "^8.0.0", - "@octokit/oauth-methods": "^6.0.1", - "@types/aws-lambda": "^8.10.83", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/oauth-app/node_modules/@octokit/auth-token": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/@octokit/auth-token/-/auth-token-6.0.0.tgz", - "integrity": "sha512-P4YJBPdPSpWTQ1NU4XYdvHvXJJDxM6YwpS0FZHRgP7YFkdVxsWcpWGy/NVqlAA7PcPCnMacXlRm1y2PFZRWL/w==", - "license": "MIT", - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/oauth-app/node_modules/@octokit/core": { - "version": "7.0.6", - "resolved": "https://registry.npmjs.org/@octokit/core/-/core-7.0.6.tgz", - "integrity": "sha512-DhGl4xMVFGVIyMwswXeyzdL4uXD5OGILGX5N8Y+f6W7LhC1Ze2poSNrkF/fedpVDHEEZ+PHFW0vL14I+mm8K3Q==", - "license": "MIT", - "dependencies": { - "@octokit/auth-token": "^6.0.0", - "@octokit/graphql": "^9.0.3", - "@octokit/request": "^10.0.6", - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0", - "before-after-hook": "^4.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/oauth-app/node_modules/@octokit/graphql": { - "version": "9.0.3", - "resolved": "https://registry.npmjs.org/@octokit/graphql/-/graphql-9.0.3.tgz", - "integrity": "sha512-grAEuupr/C1rALFnXTv6ZQhFuL1D8G5y8CN04RgrO4FIPMrtm+mcZzFG7dcBm+nq+1ppNixu+Jd78aeJOYxlGA==", - "license": "MIT", - "dependencies": { - "@octokit/request": "^10.0.6", - "@octokit/types": "^16.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/oauth-app/node_modules/before-after-hook": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/before-after-hook/-/before-after-hook-4.0.0.tgz", - "integrity": "sha512-q6tR3RPqIB1pMiTRMFcZwuG5T8vwp+vUvEG0vuI6B+Rikh5BfPp2fQ82c925FOs+b0lcFQ8CFrL+KbilfZFhOQ==", - "license": "Apache-2.0" - }, - "node_modules/@octokit/oauth-app/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, - "node_modules/@octokit/oauth-authorization-url": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/@octokit/oauth-authorization-url/-/oauth-authorization-url-8.0.0.tgz", - "integrity": "sha512-7QoLPRh/ssEA/HuHBHdVdSgF8xNLz/Bc5m9fZkArJE5bb6NmVkDm3anKxXPmN1zh6b5WKZPRr3697xKT/yM3qQ==", - "license": "MIT", - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/oauth-methods": { - "version": "6.0.2", - "resolved": "https://registry.npmjs.org/@octokit/oauth-methods/-/oauth-methods-6.0.2.tgz", - "integrity": "sha512-HiNOO3MqLxlt5Da5bZbLV8Zarnphi4y9XehrbaFMkcoJ+FL7sMxH/UlUsCVxpddVu4qvNDrBdaTVE2o4ITK8ng==", - "license": "MIT", - "dependencies": { - "@octokit/oauth-authorization-url": "^8.0.0", - "@octokit/request": "^10.0.6", - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0" - }, - "engines": { - "node": ">= 20" - } - }, "node_modules/@octokit/openapi-types": { "version": "27.0.0", "resolved": "https://registry.npmjs.org/@octokit/openapi-types/-/openapi-types-27.0.0.tgz", "integrity": "sha512-whrdktVs1h6gtR+09+QsNk2+FO+49j6ga1c55YZudfEG+oKJVvJLQi3zkOm5JjiUXAagWK2tI2kTGKJ2Ys7MGA==", - "license": "MIT" - }, - "node_modules/@octokit/openapi-webhooks-types": { - "version": "12.0.3", - "resolved": "https://registry.npmjs.org/@octokit/openapi-webhooks-types/-/openapi-webhooks-types-12.0.3.tgz", - "integrity": "sha512-90MF5LVHjBedwoHyJsgmaFhEN1uzXyBDRLEBe7jlTYx/fEhPAk3P3DAJsfZwC54m8hAIryosJOL+UuZHB3K3yA==", + "dev": true, "license": "MIT" }, "node_modules/@octokit/plugin-request-log": { @@ -2298,72 +1986,16 @@ "@octokit/openapi-types": "^24.2.0" } }, - "node_modules/@octokit/request": { - "version": "10.0.6", - "resolved": "https://registry.npmjs.org/@octokit/request/-/request-10.0.6.tgz", - "integrity": "sha512-FO+UgZCUu+pPnZAR+iKdUt64kPE7QW7ciqpldaMXaNzixz5Jld8dJ31LAUewk0cfSRkNSRKyqG438ba9c/qDlQ==", - "license": "MIT", - "dependencies": { - "@octokit/endpoint": "^11.0.2", - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0", - "fast-content-type-parse": "^3.0.0", - "universal-user-agent": "^7.0.2" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/request-error": { - "version": "7.0.2", - "resolved": "https://registry.npmjs.org/@octokit/request-error/-/request-error-7.0.2.tgz", - "integrity": "sha512-U8piOROoQQUyExw5c6dTkU3GKxts5/ERRThIauNL7yaRoeXW0q/5bgHWT7JfWBw1UyrbK8ERId2wVkcB32n0uQ==", - "license": "MIT", - "dependencies": { - "@octokit/types": "^16.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/request/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, "node_modules/@octokit/types": { "version": "16.0.0", "resolved": "https://registry.npmjs.org/@octokit/types/-/types-16.0.0.tgz", "integrity": "sha512-sKq+9r1Mm4efXW1FCk7hFSeJo4QKreL/tTbR0rz/qx/r1Oa2VV83LTA/H/MuCOX7uCIJmQVRKBcbmWoySjAnSg==", + "dev": true, "license": "MIT", "dependencies": { "@octokit/openapi-types": "^27.0.0" } }, - "node_modules/@octokit/webhooks": { - "version": "14.1.3", - "resolved": "https://registry.npmjs.org/@octokit/webhooks/-/webhooks-14.1.3.tgz", - "integrity": "sha512-gcK4FNaROM9NjA0mvyfXl0KPusk7a1BeA8ITlYEZVQCXF5gcETTd4yhAU0Kjzd8mXwYHppzJBWgdBVpIR9wUcQ==", - "license": "MIT", - "dependencies": { - "@octokit/openapi-webhooks-types": "12.0.3", - "@octokit/request-error": "^7.0.0", - "@octokit/webhooks-methods": "^6.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/@octokit/webhooks-methods": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/@octokit/webhooks-methods/-/webhooks-methods-6.0.0.tgz", - "integrity": "sha512-MFlzzoDJVw/GcbfzVC1RLR36QqkTLUf79vLVO3D+xn7r0QgxnFoLZgtrzxiQErAjFUOdH6fas2KeQJ1yr/qaXQ==", - "license": "MIT", - "engines": { - "node": ">= 20" - } - }, "node_modules/@open-draft/deferred-promise": { "version": "2.2.0", "resolved": "https://registry.npmjs.org/@open-draft/deferred-promise/-/deferred-promise-2.2.0.tgz", @@ -2398,16 +2030,6 @@ "node": ">=8.0.0" } }, - "node_modules/@pkgjs/parseargs": { - "version": "0.11.0", - "resolved": "https://registry.npmjs.org/@pkgjs/parseargs/-/parseargs-0.11.0.tgz", - "integrity": "sha512-+1VkjdD0QBLPodGrJUeqarH8VAIvQODIbwh9XpP5Syisf7YoQgsJKPNFoqqLQlu+VQ/tVSshMR6loPMn8U+dPg==", - "license": "MIT", - "optional": true, - "engines": { - "node": ">=14" - } - }, "node_modules/@pkgr/core": { "version": "0.1.1", "resolved": "https://registry.npmjs.org/@pkgr/core/-/core-0.1.1.tgz", @@ -2622,12 +2244,6 @@ "@types/readdir-glob": "*" } }, - "node_modules/@types/aws-lambda": { - "version": "8.10.157", - "resolved": "https://registry.npmjs.org/@types/aws-lambda/-/aws-lambda-8.10.157.tgz", - "integrity": "sha512-ofjcRCO1N7tMZDSO11u5bFHPDfUFD3Q9YK9g4S4w8UDKuG3CNlw2lNK1sd3Itdo7JORygZmG4h9ZykS8dlXvMA==", - "license": "MIT" - }, "node_modules/@types/color-name": { "version": "1.1.1", "dev": true, @@ -2729,9 +2345,9 @@ "license": "MIT" }, "node_modules/@types/sinon": { - "version": "17.0.4", - "resolved": "https://registry.npmjs.org/@types/sinon/-/sinon-17.0.4.tgz", - "integrity": "sha512-RHnIrhfPO3+tJT0s7cFaXGZvsL4bbR3/k7z3P312qMS4JaS2Tk+KiwiLx1S0rQ56ERj00u1/BtdyVd0FY+Pdew==", + "version": "21.0.0", + "resolved": "https://registry.npmjs.org/@types/sinon/-/sinon-21.0.0.tgz", + "integrity": "sha512-+oHKZ0lTI+WVLxx1IbJDNmReQaIsQJjN2e7UUrJHEeByG7bFeKJYsv1E75JxTQ9QKJDp21bAa/0W2Xo4srsDnw==", "dev": true, "license": "MIT", "dependencies": { @@ -3512,93 +3128,6 @@ "node": ">=18" } }, - "node_modules/@vercel/nft/node_modules/brace-expansion": { - "version": "2.0.2", - "resolved": "https://registry.npmjs.org/brace-expansion/-/brace-expansion-2.0.2.tgz", - "integrity": "sha512-Jt0vHyM+jmUBqojB7E1NIYadt0vI0Qxjxd2TErW94wDz+E2LAm5vKMXXwg6ZZBTHPuUlDgQHKXvjGBdfcF1ZDQ==", - "dev": true, - "license": "MIT", - "dependencies": { - "balanced-match": "^1.0.0" - } - }, - "node_modules/@vercel/nft/node_modules/glob": { - "version": "10.4.5", - "resolved": "https://registry.npmjs.org/glob/-/glob-10.4.5.tgz", - "integrity": "sha512-7Bv8RF0k6xjo7d4A/PxYLbUCfb6c+Vpd2/mB2yRDlew7Jb5hEXiCD9ibfO7wpk8i4sevK6DFny9h7EYbM3/sHg==", - "dev": true, - "license": "ISC", - "dependencies": { - "foreground-child": "^3.1.0", - "jackspeak": "^3.1.2", - "minimatch": "^9.0.4", - "minipass": "^7.1.2", - "package-json-from-dist": "^1.0.0", - "path-scurry": "^1.11.1" - }, - "bin": { - "glob": "dist/esm/bin.mjs" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, - "node_modules/@vercel/nft/node_modules/jackspeak": { - "version": "3.4.3", - "resolved": "https://registry.npmjs.org/jackspeak/-/jackspeak-3.4.3.tgz", - "integrity": "sha512-OGlZQpz2yfahA/Rd1Y8Cd9SIEsqvXkLVoSw/cgwhnhFMDbsQFeZYoJJ7bIZBS9BcamUW96asq/npPWugM+RQBw==", - "dev": true, - "license": "BlueOak-1.0.0", - "dependencies": { - "@isaacs/cliui": "^8.0.2" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - }, - "optionalDependencies": { - "@pkgjs/parseargs": "^0.11.0" - } - }, - "node_modules/@vercel/nft/node_modules/lru-cache": { - "version": "10.4.3", - "resolved": "https://registry.npmjs.org/lru-cache/-/lru-cache-10.4.3.tgz", - "integrity": "sha512-JNAzZcXrCt42VGLuYz0zfAzDfAvJWW6AfYlDBQyDV5DClI2m5sAmK+OIO7s59XfsRsWHp02jAJrRadPRGTt6SQ==", - "dev": true, - "license": "ISC" - }, - "node_modules/@vercel/nft/node_modules/minimatch": { - "version": "9.0.5", - "resolved": "https://registry.npmjs.org/minimatch/-/minimatch-9.0.5.tgz", - "integrity": "sha512-G6T0ZX48xgozx7587koeX9Ys2NYy6Gmv//P89sEte9V9whIapMNF4idKxnW2QtCcLiTWlb/wfCabAtAFWhhBow==", - "dev": true, - "license": "ISC", - "dependencies": { - "brace-expansion": "^2.0.1" - }, - "engines": { - "node": ">=16 || 14 >=14.17" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, - "node_modules/@vercel/nft/node_modules/path-scurry": { - "version": "1.11.1", - "resolved": "https://registry.npmjs.org/path-scurry/-/path-scurry-1.11.1.tgz", - "integrity": "sha512-Xa4Nw17FS9ApQFJ9umLiJS4orGjm7ZzwUrwamcGQuHSzDyth9boKDaycYdDcZDuqYATXw4HFXgaqWTctW/v1HA==", - "dev": true, - "license": "BlueOak-1.0.0", - "dependencies": { - "lru-cache": "^10.2.0", - "minipass": "^5.0.0 || ^6.0.2 || ^7.0.0" - }, - "engines": { - "node": ">=16 || 14 >=14.18" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, "node_modules/@vercel/nft/node_modules/picomatch": { "version": "4.0.2", "resolved": "https://registry.npmjs.org/picomatch/-/picomatch-4.0.2.tgz", @@ -3747,35 +3276,6 @@ "node": ">= 14" } }, - "node_modules/archiver-utils/node_modules/brace-expansion": { - "version": "2.0.2", - "resolved": "https://registry.npmjs.org/brace-expansion/-/brace-expansion-2.0.2.tgz", - "integrity": "sha512-Jt0vHyM+jmUBqojB7E1NIYadt0vI0Qxjxd2TErW94wDz+E2LAm5vKMXXwg6ZZBTHPuUlDgQHKXvjGBdfcF1ZDQ==", - "license": "MIT", - "dependencies": { - "balanced-match": "^1.0.0" - } - }, - "node_modules/archiver-utils/node_modules/glob": { - "version": "10.4.5", - "resolved": "https://registry.npmjs.org/glob/-/glob-10.4.5.tgz", - "integrity": "sha512-7Bv8RF0k6xjo7d4A/PxYLbUCfb6c+Vpd2/mB2yRDlew7Jb5hEXiCD9ibfO7wpk8i4sevK6DFny9h7EYbM3/sHg==", - "license": "ISC", - "dependencies": { - "foreground-child": "^3.1.0", - "jackspeak": "^3.1.2", - "minimatch": "^9.0.4", - "minipass": "^7.1.2", - "package-json-from-dist": "^1.0.0", - "path-scurry": "^1.11.1" - }, - "bin": { - "glob": "dist/esm/bin.mjs" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, "node_modules/archiver-utils/node_modules/is-stream": { "version": "2.0.1", "resolved": "https://registry.npmjs.org/is-stream/-/is-stream-2.0.1.tgz", @@ -3788,58 +3288,6 @@ "url": "https://github.com/sponsors/sindresorhus" } }, - "node_modules/archiver-utils/node_modules/jackspeak": { - "version": "3.4.3", - "resolved": "https://registry.npmjs.org/jackspeak/-/jackspeak-3.4.3.tgz", - "integrity": "sha512-OGlZQpz2yfahA/Rd1Y8Cd9SIEsqvXkLVoSw/cgwhnhFMDbsQFeZYoJJ7bIZBS9BcamUW96asq/npPWugM+RQBw==", - "license": "BlueOak-1.0.0", - "dependencies": { - "@isaacs/cliui": "^8.0.2" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - }, - "optionalDependencies": { - "@pkgjs/parseargs": "^0.11.0" - } - }, - "node_modules/archiver-utils/node_modules/lru-cache": { - "version": "10.4.3", - "resolved": "https://registry.npmjs.org/lru-cache/-/lru-cache-10.4.3.tgz", - "integrity": "sha512-JNAzZcXrCt42VGLuYz0zfAzDfAvJWW6AfYlDBQyDV5DClI2m5sAmK+OIO7s59XfsRsWHp02jAJrRadPRGTt6SQ==", - "license": "ISC" - }, - "node_modules/archiver-utils/node_modules/minimatch": { - "version": "9.0.5", - "resolved": "https://registry.npmjs.org/minimatch/-/minimatch-9.0.5.tgz", - "integrity": "sha512-G6T0ZX48xgozx7587koeX9Ys2NYy6Gmv//P89sEte9V9whIapMNF4idKxnW2QtCcLiTWlb/wfCabAtAFWhhBow==", - "license": "ISC", - "dependencies": { - "brace-expansion": "^2.0.1" - }, - "engines": { - "node": ">=16 || 14 >=14.17" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, - "node_modules/archiver-utils/node_modules/path-scurry": { - "version": "1.11.1", - "resolved": "https://registry.npmjs.org/path-scurry/-/path-scurry-1.11.1.tgz", - "integrity": "sha512-Xa4Nw17FS9ApQFJ9umLiJS4orGjm7ZzwUrwamcGQuHSzDyth9boKDaycYdDcZDuqYATXw4HFXgaqWTctW/v1HA==", - "license": "BlueOak-1.0.0", - "dependencies": { - "lru-cache": "^10.2.0", - "minipass": "^5.0.0 || ^6.0.2 || ^7.0.0" - }, - "engines": { - "node": ">=16 || 14 >=14.18" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, "node_modules/are-docs-informative": { "version": "0.0.2", "resolved": "https://registry.npmjs.org/are-docs-informative/-/are-docs-informative-0.0.2.tgz", @@ -5365,6 +4813,40 @@ "eslint": "^8 || ^9" } }, + "node_modules/eslint-plugin-github/node_modules/@eslint/compat": { + "version": "1.4.1", + "resolved": "https://registry.npmjs.org/@eslint/compat/-/compat-1.4.1.tgz", + "integrity": "sha512-cfO82V9zxxGBxcQDr1lfaYB7wykTa0b00mGa36FrJl7iTFd0Z2cHfEYuxcBRP/iNijCsWsEkA+jzT8hGYmv33w==", + "dev": true, + "license": "Apache-2.0", + "dependencies": { + "@eslint/core": "^0.17.0" + }, + "engines": { + "node": "^18.18.0 || ^20.9.0 || >=21.1.0" + }, + "peerDependencies": { + "eslint": "^8.40 || 9" + }, + "peerDependenciesMeta": { + "eslint": { + "optional": true + } + } + }, + "node_modules/eslint-plugin-github/node_modules/@eslint/core": { + "version": "0.17.0", + "resolved": "https://registry.npmjs.org/@eslint/core/-/core-0.17.0.tgz", + "integrity": "sha512-yL/sLrpmtDaFEiUj1osRP4TI2MDz1AddJL+jZ7KSqvBuliN4xqYY54IfdN8qD8Toa6g1iloph1fxQNkjOxrrpQ==", + "dev": true, + "license": "Apache-2.0", + "dependencies": { + "@types/json-schema": "^7.0.15" + }, + "engines": { + "node": "^18.18.0 || ^20.9.0 || >=21.1.0" + } + }, "node_modules/eslint-plugin-github/node_modules/debug": { "version": "3.2.7", "resolved": "https://registry.npmjs.org/debug/-/debug-3.2.7.tgz", @@ -5470,9 +4952,9 @@ } }, "node_modules/eslint-plugin-jsdoc": { - "version": "61.1.12", - "resolved": "https://registry.npmjs.org/eslint-plugin-jsdoc/-/eslint-plugin-jsdoc-61.1.12.tgz", - "integrity": "sha512-CGJTnltz7ovwOW33xYhvA4fMuriPZpR5OnJf09SV28iU2IUpJwMd6P7zvUK8Sl56u5YzO+1F9m46wpSs2dufEw==", + "version": "61.2.1", + "resolved": "https://registry.npmjs.org/eslint-plugin-jsdoc/-/eslint-plugin-jsdoc-61.2.1.tgz", + "integrity": "sha512-Htacti3dbkNm4rlp/Bk9lqhv+gi6US9jyN22yaJ42G6wbteiTbNLChQwi25jr/BN+NOzDWhZHvCDdrhX0F8dXQ==", "dev": true, "license": "BSD-3-Clause", "dependencies": { @@ -5978,22 +5460,6 @@ "url": "https://github.com/sindresorhus/execa?sponsor=1" } }, - "node_modules/fast-content-type-parse": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/fast-content-type-parse/-/fast-content-type-parse-3.0.0.tgz", - "integrity": "sha512-ZvLdcY8P+N8mGQJahJV5G4U88CSvT1rP8ApL6uETe88MBXrBHAkZlSEySdUlyztF7ccb+Znos3TFqaepHxdhBg==", - "funding": [ - { - "type": "github", - "url": "https://github.com/sponsors/fastify" - }, - { - "type": "opencollective", - "url": "https://opencollective.com/fastify" - } - ], - "license": "MIT" - }, "node_modules/fast-deep-equal": { "version": "3.1.3", "license": "MIT" @@ -6198,11 +5664,6 @@ "node": ">= 0.12" } }, - "node_modules/fs.realpath": { - "version": "1.0.0", - "dev": true, - "license": "ISC" - }, "node_modules/function-bind": { "version": "1.1.2", "license": "MIT", @@ -6351,15 +5812,14 @@ } }, "node_modules/glob": { - "version": "11.0.3", - "resolved": "https://registry.npmjs.org/glob/-/glob-11.0.3.tgz", - "integrity": "sha512-2Nim7dha1KVkaiF4q6Dj+ngPPMdfvLJEOpZk/jKiUAkqKebpGAWQXAq9z1xu9HKu5lWfqw/FASuccEjyznjPaA==", - "dev": true, - "license": "ISC", + "version": "11.1.0", + "resolved": "https://registry.npmjs.org/glob/-/glob-11.1.0.tgz", + "integrity": "sha512-vuNwKSaKiqm7g0THUBu2x7ckSs3XJLXE+2ssL7/MfTGPLLcrJQ/4Uq1CjPTtO5cCIiRxqvN6Twy1qOwhL0Xjcw==", + "license": "BlueOak-1.0.0", "dependencies": { "foreground-child": "^3.3.1", "jackspeak": "^4.1.1", - "minimatch": "^10.0.3", + "minimatch": "^10.1.1", "minipass": "^7.1.2", "package-json-from-dist": "^1.0.0", "path-scurry": "^2.0.0" @@ -6386,11 +5846,10 @@ } }, "node_modules/glob/node_modules/minimatch": { - "version": "10.0.3", - "resolved": "https://registry.npmjs.org/minimatch/-/minimatch-10.0.3.tgz", - "integrity": "sha512-IPZ167aShDZZUMdRk66cyQAW3qr0WzbHkPdMYa8bzZhlHhO3jALbKdxcaak7W9FfT2rZNpQuUu4Od7ILEpXSaw==", - "dev": true, - "license": "ISC", + "version": "10.1.1", + "resolved": "https://registry.npmjs.org/minimatch/-/minimatch-10.1.1.tgz", + "integrity": "sha512-enIvLvRAFZYXJzkCYG5RKmPfrFArdLv+R+lbQ53BmIMLIry74bjKzX6iHAm8WYamJkhSSEabrWN5D97XnKObjQ==", + "license": "BlueOak-1.0.0", "dependencies": { "@isaacs/brace-expansion": "^5.0.0" }, @@ -6702,15 +6161,6 @@ "url": "https://github.com/sponsors/sindresorhus" } }, - "node_modules/inflight": { - "version": "1.0.6", - "dev": true, - "license": "ISC", - "dependencies": { - "once": "^1.3.0", - "wrappy": "1" - } - }, "node_modules/inherits": { "version": "2.0.3", "license": "ISC" @@ -7067,7 +6517,6 @@ "version": "4.1.1", "resolved": "https://registry.npmjs.org/jackspeak/-/jackspeak-4.1.1.tgz", "integrity": "sha512-zptv57P3GpL+O0I7VdMJNBZCu+BPHVQUk55Ft8/QCJjTVxrnJHuVuX/0Bl2A6/+2oyR/ZMEuFKwmzqqZ/U5nPQ==", - "dev": true, "license": "BlueOak-1.0.0", "dependencies": { "@isaacs/cliui": "^8.0.2" @@ -7089,7 +6538,9 @@ } }, "node_modules/js-yaml": { - "version": "4.1.0", + "version": "4.1.1", + "resolved": "https://registry.npmjs.org/js-yaml/-/js-yaml-4.1.1.tgz", + "integrity": "sha512-qQKT4zQxXl8lLwBtHMWwaTcGfFOZviOJet3Oy/xmGk2gZH677CJM9EvtfdSkgWcATZhj/55JZ0rmy3myCT5lsA==", "license": "MIT", "dependencies": { "argparse": "^2.0.1" @@ -7099,10 +6550,11 @@ } }, "node_modules/jschardet": { - "version": "3.1.3", - "resolved": "https://registry.npmjs.org/jschardet/-/jschardet-3.1.3.tgz", - "integrity": "sha512-Q1PKVMK/uu+yjdlobgWIYkUOCR1SqUmW9m/eUJNNj4zI2N12i25v8fYpVf+zCakQeaTdBdhnZTFbVIAVZIVVOg==", + "version": "3.1.4", + "resolved": "https://registry.npmjs.org/jschardet/-/jschardet-3.1.4.tgz", + "integrity": "sha512-/kmVISmrwVwtyYU40iQUOp3SUPk2dhNCMsZBQX0R1/jZ8maaXJ/oZIzUOiyOqcgtLnETFKYChbJ5iDC/eWmFHg==", "dev": true, + "license": "LGPL-2.1+", "engines": { "node": ">=0.1.90" } @@ -7308,7 +6760,6 @@ "version": "11.1.0", "resolved": "https://registry.npmjs.org/lru-cache/-/lru-cache-11.1.0.tgz", "integrity": "sha512-QIXZUBJUx+2zHUdQujWejBkcD9+cs94tLn0+YL8UrCh+D5sCXZ4c7LaEH48pNwRY3MLDgqUFyhlCyjJPf1WP0A==", - "dev": true, "license": "ISC", "engines": { "node": "20 || >=22" @@ -7699,153 +7150,6 @@ "url": "https://github.com/sponsors/ljharb" } }, - "node_modules/octokit": { - "version": "5.0.5", - "resolved": "https://registry.npmjs.org/octokit/-/octokit-5.0.5.tgz", - "integrity": "sha512-4+/OFSqOjoyULo7eN7EA97DE0Xydj/PW5aIckxqQIoFjFwqXKuFCvXUJObyJfBF9Khu4RL/jlDRI9FPaMGfPnw==", - "license": "MIT", - "dependencies": { - "@octokit/app": "^16.1.2", - "@octokit/core": "^7.0.6", - "@octokit/oauth-app": "^8.0.3", - "@octokit/plugin-paginate-graphql": "^6.0.0", - "@octokit/plugin-paginate-rest": "^14.0.0", - "@octokit/plugin-rest-endpoint-methods": "^17.0.0", - "@octokit/plugin-retry": "^8.0.3", - "@octokit/plugin-throttling": "^11.0.3", - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0", - "@octokit/webhooks": "^14.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/octokit/node_modules/@octokit/auth-token": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/@octokit/auth-token/-/auth-token-6.0.0.tgz", - "integrity": "sha512-P4YJBPdPSpWTQ1NU4XYdvHvXJJDxM6YwpS0FZHRgP7YFkdVxsWcpWGy/NVqlAA7PcPCnMacXlRm1y2PFZRWL/w==", - "engines": { - "node": ">= 20" - } - }, - "node_modules/octokit/node_modules/@octokit/core": { - "version": "7.0.6", - "resolved": "https://registry.npmjs.org/@octokit/core/-/core-7.0.6.tgz", - "integrity": "sha512-DhGl4xMVFGVIyMwswXeyzdL4uXD5OGILGX5N8Y+f6W7LhC1Ze2poSNrkF/fedpVDHEEZ+PHFW0vL14I+mm8K3Q==", - "license": "MIT", - "dependencies": { - "@octokit/auth-token": "^6.0.0", - "@octokit/graphql": "^9.0.3", - "@octokit/request": "^10.0.6", - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0", - "before-after-hook": "^4.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/octokit/node_modules/@octokit/graphql": { - "version": "9.0.3", - "resolved": "https://registry.npmjs.org/@octokit/graphql/-/graphql-9.0.3.tgz", - "integrity": "sha512-grAEuupr/C1rALFnXTv6ZQhFuL1D8G5y8CN04RgrO4FIPMrtm+mcZzFG7dcBm+nq+1ppNixu+Jd78aeJOYxlGA==", - "license": "MIT", - "dependencies": { - "@octokit/request": "^10.0.6", - "@octokit/types": "^16.0.0", - "universal-user-agent": "^7.0.0" - }, - "engines": { - "node": ">= 20" - } - }, - "node_modules/octokit/node_modules/@octokit/plugin-paginate-graphql": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/@octokit/plugin-paginate-graphql/-/plugin-paginate-graphql-6.0.0.tgz", - "integrity": "sha512-crfpnIoFiBtRkvPqOyLOsw12XsveYuY2ieP6uYDosoUegBJpSVxGwut9sxUgFFcll3VTOTqpUf8yGd8x1OmAkQ==", - "engines": { - "node": ">= 20" - }, - "peerDependencies": { - "@octokit/core": ">=6" - } - }, - "node_modules/octokit/node_modules/@octokit/plugin-paginate-rest": { - "version": "14.0.0", - "resolved": "https://registry.npmjs.org/@octokit/plugin-paginate-rest/-/plugin-paginate-rest-14.0.0.tgz", - "integrity": "sha512-fNVRE7ufJiAA3XUrha2omTA39M6IXIc6GIZLvlbsm8QOQCYvpq/LkMNGyFlB1d8hTDzsAXa3OKtybdMAYsV/fw==", - "license": "MIT", - "dependencies": { - "@octokit/types": "^16.0.0" - }, - "engines": { - "node": ">= 20" - }, - "peerDependencies": { - "@octokit/core": ">=6" - } - }, - "node_modules/octokit/node_modules/@octokit/plugin-rest-endpoint-methods": { - "version": "17.0.0", - "resolved": "https://registry.npmjs.org/@octokit/plugin-rest-endpoint-methods/-/plugin-rest-endpoint-methods-17.0.0.tgz", - "integrity": "sha512-B5yCyIlOJFPqUUeiD0cnBJwWJO8lkJs5d8+ze9QDP6SvfiXSz1BF+91+0MeI1d2yxgOhU/O+CvtiZ9jSkHhFAw==", - "license": "MIT", - "dependencies": { - "@octokit/types": "^16.0.0" - }, - "engines": { - "node": ">= 20" - }, - "peerDependencies": { - "@octokit/core": ">=6" - } - }, - "node_modules/octokit/node_modules/@octokit/plugin-retry": { - "version": "8.0.3", - "resolved": "https://registry.npmjs.org/@octokit/plugin-retry/-/plugin-retry-8.0.3.tgz", - "integrity": "sha512-vKGx1i3MC0za53IzYBSBXcrhmd+daQDzuZfYDd52X5S0M2otf3kVZTVP8bLA3EkU0lTvd1WEC2OlNNa4G+dohA==", - "license": "MIT", - "dependencies": { - "@octokit/request-error": "^7.0.2", - "@octokit/types": "^16.0.0", - "bottleneck": "^2.15.3" - }, - "engines": { - "node": ">= 20" - }, - "peerDependencies": { - "@octokit/core": ">=7" - } - }, - "node_modules/octokit/node_modules/@octokit/plugin-throttling": { - "version": "11.0.3", - "resolved": "https://registry.npmjs.org/@octokit/plugin-throttling/-/plugin-throttling-11.0.3.tgz", - "integrity": "sha512-34eE0RkFCKycLl2D2kq7W+LovheM/ex3AwZCYN8udpi6bxsyjZidb2McXs69hZhLmJlDqTSP8cH+jSRpiaijBg==", - "license": "MIT", - "dependencies": { - "@octokit/types": "^16.0.0", - "bottleneck": "^2.15.3" - }, - "engines": { - "node": ">= 20" - }, - "peerDependencies": { - "@octokit/core": "^7.0.0" - } - }, - "node_modules/octokit/node_modules/before-after-hook": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/before-after-hook/-/before-after-hook-4.0.0.tgz", - "integrity": "sha512-q6tR3RPqIB1pMiTRMFcZwuG5T8vwp+vUvEG0vuI6B+Rikh5BfPp2fQ82c925FOs+b0lcFQ8CFrL+KbilfZFhOQ==" - }, - "node_modules/octokit/node_modules/universal-user-agent": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-7.0.3.tgz", - "integrity": "sha512-TmnEAEAsBJVZM/AADELsK76llnwcf9vMKuPz8JflO1frO8Lchitr0fNaN9d+Ap0BjKtqWqd/J17qeDnXh8CL2A==", - "license": "ISC" - }, "node_modules/once": { "version": "1.4.0", "license": "ISC", @@ -7950,14 +7254,6 @@ "dev": true, "license": "MIT" }, - "node_modules/path-is-absolute": { - "version": "1.0.1", - "dev": true, - "license": "MIT", - "engines": { - "node": ">=0.10.0" - } - }, "node_modules/path-key": { "version": "3.1.1", "license": "MIT", @@ -7974,7 +7270,6 @@ "version": "2.0.0", "resolved": "https://registry.npmjs.org/path-scurry/-/path-scurry-2.0.0.tgz", "integrity": "sha512-ypGJsmGtdXUOeM5u93TyeIEfEhM6s+ljAhrk5vAvSx8uyY/02OvrZnA0YNGUrPXfpJMgI1ODd3nwz8Npx4O4cg==", - "dev": true, "license": "BlueOak-1.0.0", "dependencies": { "lru-cache": "^11.0.0", @@ -8312,25 +7607,6 @@ "url": "https://github.com/sponsors/isaacs" } }, - "node_modules/rimraf/node_modules/glob": { - "version": "7.2.0", - "dev": true, - "license": "ISC", - "dependencies": { - "fs.realpath": "^1.0.0", - "inflight": "^1.0.4", - "inherits": "2", - "minimatch": "^3.0.4", - "once": "^1.3.0", - "path-is-absolute": "^1.0.0" - }, - "engines": { - "node": "*" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, "node_modules/run-parallel": { "version": "1.2.0", "dev": true, @@ -9117,15 +8393,6 @@ "url": "https://github.com/sponsors/sindresorhus" } }, - "node_modules/toad-cache": { - "version": "3.7.0", - "resolved": "https://registry.npmjs.org/toad-cache/-/toad-cache-3.7.0.tgz", - "integrity": "sha512-/m8M+2BJUpoJdgAHoG+baCwBT+tf2VraSfkBgl0Y00qIWt41DJ8R5B8nsEw0I58YwF5IZH6z24/2TobDKnqSWw==", - "license": "MIT", - "engines": { - "node": ">=12" - } - }, "node_modules/tr46": { "version": "0.0.3", "license": "MIT" @@ -9448,12 +8715,6 @@ "url": "https://github.com/sponsors/sindresorhus" } }, - "node_modules/universal-github-app-jwt": { - "version": "2.2.2", - "resolved": "https://registry.npmjs.org/universal-github-app-jwt/-/universal-github-app-jwt-2.2.2.tgz", - "integrity": "sha512-dcmbeSrOdTnsjGjUfAlqNDJrhxXizjAz94ija9Qw8YkZ1uu0d+GoZzyH+Jb9tIIqvGsadUfwg+22k5aDqqwzbw==", - "license": "MIT" - }, "node_modules/universal-user-agent": { "version": "6.0.0", "license": "ISC" @@ -9511,7 +8772,8 @@ "version": "3.0.0", "resolved": "https://registry.npmjs.org/utf8/-/utf8-3.0.0.tgz", "integrity": "sha512-E8VjFIQ/TyQgp+TZfS6l8yp/xWppSAHzidGiRrqe4bK4XP9pTRyKFgGJpO3SN7zdX4DeomTrwaseCHovfpFcqQ==", - "dev": true + "dev": true, + "license": "MIT" }, "node_modules/util-deprecate": { "version": "1.0.2", diff --git a/package.json b/package.json index dcad231e2..22d581776 100644 --- a/package.json +++ b/package.json @@ -1,6 +1,6 @@ { "name": "codeql", - "version": "4.31.3", + "version": "4.31.6", "private": true, "description": "CodeQL action", "scripts": { @@ -35,23 +35,21 @@ "@actions/io": "^2.0.0", "@actions/tool-cache": "^2.0.2", "@octokit/plugin-retry": "^6.0.0", - "@octokit/request-error": "^7.0.2", "@schemastore/package": "0.0.10", "archiver": "^7.0.1", "fast-deep-equal": "^3.1.3", "follow-redirects": "^1.15.11", "get-folder-size": "^5.0.0", - "js-yaml": "^4.1.0", + "js-yaml": "^4.1.1", "jsonschema": "1.4.1", "long": "^5.3.2", "node-forge": "^1.3.1", - "octokit": "^5.0.5", "semver": "^7.7.3", "uuid": "^13.0.0" }, "devDependencies": { "@ava/typescript": "6.0.0", - "@eslint/compat": "^1.4.1", + "@eslint/compat": "^2.0.0", "@eslint/eslintrc": "^3.3.1", "@eslint/js": "^9.39.1", "@microsoft/eslint-formatter-sarif": "^3.1.0", @@ -59,10 +57,10 @@ "@types/archiver": "^7.0.0", "@types/follow-redirects": "^1.14.4", "@types/js-yaml": "^4.0.9", - "@types/node": "20.19.9", + "@types/node": "^20.19.9", "@types/node-forge": "^1.3.14", "@types/semver": "^7.7.1", - "@types/sinon": "^17.0.4", + "@types/sinon": "^21.0.0", "@typescript-eslint/eslint-plugin": "^8.46.4", "@typescript-eslint/parser": "^8.41.0", "ava": "^6.4.1", @@ -72,9 +70,9 @@ "eslint-plugin-filenames": "^1.3.2", "eslint-plugin-github": "^5.1.8", "eslint-plugin-import": "2.29.1", - "eslint-plugin-jsdoc": "^61.1.12", + "eslint-plugin-jsdoc": "^61.2.1", "eslint-plugin-no-async-foreach": "^0.1.1", - "glob": "^11.0.3", + "glob": "^11.1.0", "nock": "^14.0.10", "sinon": "^21.0.0", "typescript": "^5.9.3" @@ -98,6 +96,7 @@ "eslint-plugin-jsx-a11y": { "semver": ">=6.3.1" }, - "brace-expansion@2.0.1": "2.0.2" + "brace-expansion@2.0.1": "2.0.2", + "glob": "^11.1.0" } } diff --git a/pr-checks/checks/all-platform-bundle.yml b/pr-checks/checks/all-platform-bundle.yml index 3396be22a..994c91eb9 100644 --- a/pr-checks/checks/all-platform-bundle.yml +++ b/pr-checks/checks/all-platform-bundle.yml @@ -4,6 +4,7 @@ operatingSystems: ["ubuntu", "macos", "windows"] versions: ["nightly-latest"] useAllPlatformBundle: "true" installGo: true +installDotNet: true steps: - id: init uses: ./../action/init diff --git a/pr-checks/checks/analyze-ref-input.yml b/pr-checks/checks/analyze-ref-input.yml index 1dbf9e0c8..e9d2cd176 100644 --- a/pr-checks/checks/analyze-ref-input.yml +++ b/pr-checks/checks/analyze-ref-input.yml @@ -3,6 +3,7 @@ description: "Checks that specifying 'ref' and 'sha' as inputs works" versions: ["default"] installGo: true installPython: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/autobuild-action.yml b/pr-checks/checks/autobuild-action.yml index 91ae7834c..b91489cc8 100644 --- a/pr-checks/checks/autobuild-action.yml +++ b/pr-checks/checks/autobuild-action.yml @@ -2,6 +2,7 @@ name: "autobuild-action" description: "Tests that the C# autobuild action works" operatingSystems: ["ubuntu", "macos", "windows"] versions: ["linked"] +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/build-mode-manual.yml b/pr-checks/checks/build-mode-manual.yml index f1815b7ff..a8048230c 100644 --- a/pr-checks/checks/build-mode-manual.yml +++ b/pr-checks/checks/build-mode-manual.yml @@ -2,6 +2,7 @@ name: "Build mode manual" description: "An end-to-end integration test of a Java repository built using 'build-mode: manual'" versions: ["nightly-latest"] installGo: true +installDotNet: true steps: - uses: ./../action/init id: init diff --git a/pr-checks/checks/export-file-baseline-information.yml b/pr-checks/checks/export-file-baseline-information.yml index 1a316d6ec..5ff186e9f 100644 --- a/pr-checks/checks/export-file-baseline-information.yml +++ b/pr-checks/checks/export-file-baseline-information.yml @@ -3,6 +3,7 @@ description: "Tests that file baseline information is exported when the feature operatingSystems: ["ubuntu", "macos", "windows"] versions: ["nightly-latest"] installGo: true +installDotNet: true env: CODEQL_ACTION_SUBLANGUAGE_FILE_COVERAGE: true steps: diff --git a/pr-checks/checks/go-custom-queries.yml b/pr-checks/checks/go-custom-queries.yml index ca00fd81a..867d8dc6c 100644 --- a/pr-checks/checks/go-custom-queries.yml +++ b/pr-checks/checks/go-custom-queries.yml @@ -7,6 +7,7 @@ versions: - linked - nightly-latest installGo: true +installDotNet: true env: DOTNET_GENERATE_ASPNET_CERTIFICATE: "false" steps: diff --git a/pr-checks/checks/local-bundle.yml b/pr-checks/checks/local-bundle.yml index 449dcbbb1..c0930772e 100644 --- a/pr-checks/checks/local-bundle.yml +++ b/pr-checks/checks/local-bundle.yml @@ -3,6 +3,7 @@ description: "Tests using a CodeQL bundle from a local file rather than a URL" versions: ["linked"] installGo: true installPython: true +installDotNet: true steps: - name: Fetch latest CodeQL bundle run: | diff --git a/pr-checks/checks/multi-language-autodetect.yml b/pr-checks/checks/multi-language-autodetect.yml index 84c779dd9..4892bcc31 100644 --- a/pr-checks/checks/multi-language-autodetect.yml +++ b/pr-checks/checks/multi-language-autodetect.yml @@ -5,6 +5,7 @@ env: CODEQL_ACTION_RESOLVE_SUPPORTED_LANGUAGES_USING_CLI: true installGo: true installPython: true +installDotNet: true steps: - name: Use Xcode 16 if: runner.os == 'macOS' && matrix.version != 'nightly-latest' diff --git a/pr-checks/checks/packaging-codescanning-config-inputs-js.yml b/pr-checks/checks/packaging-codescanning-config-inputs-js.yml index 1b72b06b3..6fd0f7c8a 100644 --- a/pr-checks/checks/packaging-codescanning-config-inputs-js.yml +++ b/pr-checks/checks/packaging-codescanning-config-inputs-js.yml @@ -4,6 +4,7 @@ versions: ["linked", "default", "nightly-latest"] # This feature is not compatib installGo: true installNode: true installPython: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/packaging-config-inputs-js.yml b/pr-checks/checks/packaging-config-inputs-js.yml index 41275fd15..8df42f944 100644 --- a/pr-checks/checks/packaging-config-inputs-js.yml +++ b/pr-checks/checks/packaging-config-inputs-js.yml @@ -3,6 +3,7 @@ description: "Checks that specifying packages using a combination of a config fi versions: ["linked", "default", "nightly-latest"] # This feature is not compatible with old CLIs installGo: true installNode: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/packaging-config-js.yml b/pr-checks/checks/packaging-config-js.yml index 906a3a7d9..9fa41061c 100644 --- a/pr-checks/checks/packaging-config-js.yml +++ b/pr-checks/checks/packaging-config-js.yml @@ -3,6 +3,7 @@ description: "Checks that specifying packages using only a config file works" versions: ["linked", "default", "nightly-latest"] # This feature is not compatible with old CLIs installGo: true installNode: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/packaging-inputs-js.yml b/pr-checks/checks/packaging-inputs-js.yml index 9d9fbe71f..bb70de7e6 100644 --- a/pr-checks/checks/packaging-inputs-js.yml +++ b/pr-checks/checks/packaging-inputs-js.yml @@ -3,6 +3,7 @@ description: "Checks that specifying packages using the input to the Action work versions: ["linked", "default", "nightly-latest"] # This feature is not compatible with old CLIs installGo: true installNode: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/remote-config.yml b/pr-checks/checks/remote-config.yml index c0a9cf0bb..24249156e 100644 --- a/pr-checks/checks/remote-config.yml +++ b/pr-checks/checks/remote-config.yml @@ -7,6 +7,7 @@ versions: - nightly-latest installGo: true installPython: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/rubocop-multi-language.yml b/pr-checks/checks/rubocop-multi-language.yml index 8a1f91444..192660816 100644 --- a/pr-checks/checks/rubocop-multi-language.yml +++ b/pr-checks/checks/rubocop-multi-language.yml @@ -4,7 +4,7 @@ description: "Tests using RuboCop to analyze a multi-language repository and the versions: ["default"] steps: - name: Set up Ruby - uses: ruby/setup-ruby@d5126b9b3579e429dd52e51e68624dda2e05be25 # v1.267.0 + uses: ruby/setup-ruby@8aeb6ff8030dd539317f8e1769a044873b56ea71 # v1.268.0 with: ruby-version: 2.6 - name: Install Code Scanning integration diff --git a/pr-checks/checks/split-workflow.yml b/pr-checks/checks/split-workflow.yml index fdcf1d530..23f82a7a5 100644 --- a/pr-checks/checks/split-workflow.yml +++ b/pr-checks/checks/split-workflow.yml @@ -3,6 +3,7 @@ description: "Tests a split-up workflow in which we first build a database and l operatingSystems: ["ubuntu", "macos"] versions: ["linked", "default", "nightly-latest"] # This feature is not compatible with old CLIs installGo: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/submit-sarif-failure.yml b/pr-checks/checks/submit-sarif-failure.yml index 97332e4c9..5db63bb81 100644 --- a/pr-checks/checks/submit-sarif-failure.yml +++ b/pr-checks/checks/submit-sarif-failure.yml @@ -18,7 +18,7 @@ permissions: security-events: write # needed to upload the SARIF file steps: - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 - uses: ./init with: languages: javascript diff --git a/pr-checks/checks/swift-custom-build.yml b/pr-checks/checks/swift-custom-build.yml index 2ad44ff3b..1c8f1bf3a 100644 --- a/pr-checks/checks/swift-custom-build.yml +++ b/pr-checks/checks/swift-custom-build.yml @@ -3,6 +3,7 @@ description: "Tests creation of a Swift database using custom build" versions: ["linked", "default", "nightly-latest"] operatingSystems: ["macos"] installGo: true +installDotNet: true env: DOTNET_GENERATE_ASPNET_CERTIFICATE: "false" steps: diff --git a/pr-checks/checks/unset-environment.yml b/pr-checks/checks/unset-environment.yml index 614c2006e..4cc728600 100644 --- a/pr-checks/checks/unset-environment.yml +++ b/pr-checks/checks/unset-environment.yml @@ -7,6 +7,7 @@ versions: - nightly-latest installGo: true installPython: true +installDotNet: true steps: - uses: ./../action/init id: init diff --git a/pr-checks/checks/upload-ref-sha-input.yml b/pr-checks/checks/upload-ref-sha-input.yml index fe55b5457..0c8059a51 100644 --- a/pr-checks/checks/upload-ref-sha-input.yml +++ b/pr-checks/checks/upload-ref-sha-input.yml @@ -3,6 +3,7 @@ description: "Checks that specifying 'ref' and 'sha' as inputs works" versions: ["default"] installGo: true installPython: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/upload-sarif.yml b/pr-checks/checks/upload-sarif.yml index 0028e67ed..cfe66a3f8 100644 --- a/pr-checks/checks/upload-sarif.yml +++ b/pr-checks/checks/upload-sarif.yml @@ -4,6 +4,7 @@ versions: ["default"] analysisKinds: ["code-scanning", "code-quality", "code-scanning,code-quality"] installGo: true installPython: true +installDotNet: true steps: - uses: ./../action/init with: diff --git a/pr-checks/checks/with-checkout-path.yml b/pr-checks/checks/with-checkout-path.yml index b18d191cc..230e342e3 100644 --- a/pr-checks/checks/with-checkout-path.yml +++ b/pr-checks/checks/with-checkout-path.yml @@ -3,6 +3,7 @@ description: "Checks that a custom `checkout_path` will find the proper commit_o versions: ["linked"] installGo: true installPython: true +installDotNet: true steps: # This ensures we don't accidentally use the original checkout for any part of the test. - name: Delete original checkout @@ -13,7 +14,7 @@ steps: rm -rf ./* .github .git # Check out the actions repo again, but at a different location. # choose an arbitrary SHA so that we can later test that the commit_oid is not from main - - uses: actions/checkout@v5 + - uses: actions/checkout@v6 with: ref: 474bbf07f9247ffe1856c6a0f94aeeb10e7afee6 path: x/y/z/some-path diff --git a/pr-checks/sync.py b/pr-checks/sync.py index f247f3824..7d412e9b0 100755 --- a/pr-checks/sync.py +++ b/pr-checks/sync.py @@ -107,7 +107,7 @@ for file in sorted((this_dir / 'checks').glob('*.yml')): steps = [ { 'name': 'Check out repository', - 'uses': 'actions/checkout@v5' + 'uses': 'actions/checkout@v6' }, ] @@ -204,6 +204,25 @@ for file in sorted((this_dir / 'checks').glob('*.yml')): } }) + installDotNet = is_truthy(checkSpecification.get('installDotNet', '')) + + if installDotNet: + baseDotNetVersionExpr = '9.x' + workflowInputs['dotnet-version'] = { + 'type': 'string', + 'description': 'The version of .NET to install', + 'required': False, + 'default': baseDotNetVersionExpr, + } + + steps.append({ + 'name': 'Install .NET', + 'uses': 'actions/setup-dotnet@v5', + 'with': { + 'dotnet-version': '${{ inputs.dotnet-version || \'' + baseDotNetVersionExpr + '\' }}' + } + }) + # If container initialisation steps are present in the check specification, # make sure to execute them first. if 'container' in checkSpecification and 'container-init-steps' in checkSpecification: diff --git a/src/actions-util.ts b/src/actions-util.ts index a2d691b42..736d35d0f 100644 --- a/src/actions-util.ts +++ b/src/actions-util.ts @@ -80,7 +80,7 @@ export function isRunningLocalAction(): boolean { * * This can be used to get the Action's name or tell if we're running a local Action. */ -export function getRelativeScriptPath(): string { +function getRelativeScriptPath(): string { const runnerTemp = getRequiredEnvParam("RUNNER_TEMP"); const actionsDirectory = path.join(path.dirname(runnerTemp), "_actions"); return path.relative(actionsDirectory, __filename); diff --git a/src/analyses.ts b/src/analyses.ts index a8d172e3b..4f91ab07c 100644 --- a/src/analyses.ts +++ b/src/analyses.ts @@ -98,7 +98,7 @@ export async function getAnalysisKinds( export const codeQualityQueries: string[] = ["code-quality"]; // Enumerates API endpoints that accept SARIF files. -export enum SARIF_UPLOAD_ENDPOINT { +enum SARIF_UPLOAD_ENDPOINT { CODE_SCANNING = "PUT /repos/:owner/:repo/code-scanning/analysis", CODE_QUALITY = "PUT /repos/:owner/:repo/code-quality/analysis", } diff --git a/src/analyze-action.ts b/src/analyze-action.ts index 0349c13c3..f89eed7d1 100644 --- a/src/analyze-action.ts +++ b/src/analyze-action.ts @@ -25,7 +25,7 @@ import { isCodeQualityEnabled, isCodeScanningEnabled, } from "./config-utils"; -import { uploadDatabases } from "./database-upload"; +import { cleanupAndUploadDatabases } from "./database-upload"; import { DependencyCacheUploadStatusReport, uploadDependencyCaches, @@ -35,7 +35,7 @@ import { EnvVar } from "./environment"; import { Feature, Features } from "./feature-flags"; import { KnownLanguage } from "./languages"; import { getActionsLogger, Logger } from "./logging"; -import { uploadOverlayBaseDatabaseToCache } from "./overlay-database-utils"; +import { cleanupAndUploadOverlayBaseDatabaseToCache } from "./overlay-database-utils"; import { getRepositoryNwo } from "./repository"; import * as statusReport from "./status-report"; import { @@ -418,12 +418,21 @@ async function run() { } // Possibly upload the overlay-base database to actions cache. - // If databases are to be uploaded, they will first be cleaned up at the overlay level. - await uploadOverlayBaseDatabaseToCache(codeql, config, logger); + // Note: Take care with the ordering of this call since databases may be cleaned up + // at the `overlay` level. + await cleanupAndUploadOverlayBaseDatabaseToCache(codeql, config, logger); // Possibly upload the database bundles for remote queries. - // If databases are to be uploaded, they will first be cleaned up at the clear level. - await uploadDatabases(repositoryNwo, codeql, config, apiDetails, logger); + // Note: Take care with the ordering of this call since databases may be cleaned up + // at the `overlay` or `clear` level. + await cleanupAndUploadDatabases( + repositoryNwo, + codeql, + config, + apiDetails, + features, + logger, + ); // Possibly upload the TRAP caches for later re-use const trapCacheUploadStartTime = performance.now(); diff --git a/src/api-client.ts b/src/api-client.ts index e14048337..600da1ed6 100644 --- a/src/api-client.ts +++ b/src/api-client.ts @@ -18,11 +18,6 @@ import { const GITHUB_ENTERPRISE_VERSION_HEADER = "x-github-enterprise-version"; -export enum DisallowedAPIVersionReason { - ACTION_TOO_OLD, - ACTION_TOO_NEW, -} - export type GitHubApiCombinedDetails = GitHubApiDetails & GitHubApiExternalRepoDetails; diff --git a/src/cli-errors.ts b/src/cli-errors.ts index 0a009f983..5aba268ca 100644 --- a/src/cli-errors.ts +++ b/src/cli-errors.ts @@ -159,10 +159,7 @@ type CliErrorConfiguration = { * All of our caught CLI error messages that we handle specially: ie. if we * would like to categorize an error as a configuration error or not. */ -export const cliErrorsConfig: Record< - CliConfigErrorCategory, - CliErrorConfiguration -> = { +const cliErrorsConfig: Record = { [CliConfigErrorCategory.AutobuildError]: { cliErrorMessageCandidates: [ new RegExp("We were unable to automatically build your code"), diff --git a/src/codeql.ts b/src/codeql.ts index bc1d00401..2b86d843f 100644 --- a/src/codeql.ts +++ b/src/codeql.ts @@ -35,7 +35,7 @@ import { ToolsDownloadStatusReport } from "./tools-download"; import { ToolsFeature, isSupportedToolsFeature } from "./tools-features"; import { shouldEnableIndirectTracing } from "./tracer-config"; import * as util from "./util"; -import { BuildMode, getErrorMessage } from "./util"; +import { BuildMode, CleanupLevel, getErrorMessage } from "./util"; type Options = Array; @@ -141,7 +141,10 @@ export interface CodeQL { /** * Clean up all the databases within a database cluster. */ - databaseCleanupCluster(config: Config, cleanupLevel: string): Promise; + databaseCleanupCluster( + config: Config, + cleanupLevel: CleanupLevel, + ): Promise; /** * Run 'codeql database bundle'. */ @@ -513,7 +516,7 @@ export async function getCodeQLForTesting( * version requirement. Must be set to true outside tests. * @returns A new CodeQL object */ -export async function getCodeQLForCmd( +async function getCodeQLForCmd( cmd: string, checkVersion: boolean, ): Promise { @@ -878,7 +881,7 @@ export async function getCodeQLForCmd( }, async databaseCleanupCluster( config: Config, - cleanupLevel: string, + cleanupLevel: CleanupLevel, ): Promise { const cacheCleanupFlag = (await util.codeQlVersionAtLeast( this, @@ -1222,7 +1225,7 @@ export async function getTrapCachingExtractorConfigArgsForLang( * * This will not exist if the configuration is being parsed in the Action. */ -export function getGeneratedCodeScanningConfigPath(config: Config): string { +function getGeneratedCodeScanningConfigPath(config: Config): string { return path.resolve(config.tempDir, "user-config.yaml"); } diff --git a/src/config-utils.test.ts b/src/config-utils.test.ts index 32c794450..7f991ea24 100644 --- a/src/config-utils.test.ts +++ b/src/config-utils.test.ts @@ -37,7 +37,9 @@ import { ConfigurationError, withTmpDir, BuildMode, + DiskUsage, } from "./util"; +import * as util from "./util"; setupTests(test); @@ -200,12 +202,9 @@ test("load code quality config", async (t) => { ); // And the config we expect it to result in - const expectedConfig: configUtils.Config = { - version: actionsUtil.getActionVersion(), + const expectedConfig = createTestConfig({ analysisKinds: [AnalysisKind.CodeQuality], languages: [KnownLanguage.actions], - buildMode: undefined, - originalUserInput: {}, // This gets set because we only have `AnalysisKind.CodeQuality` computedConfig: { "disable-default-queries": true, @@ -219,14 +218,7 @@ test("load code quality config", async (t) => { debugMode: false, debugArtifactName: "", debugDatabaseName: "", - trapCaches: {}, - trapCacheDownloadTime: 0, - dependencyCachingEnabled: CachingKind.None, - extraQueryExclusions: [], - overlayDatabaseMode: OverlayDatabaseMode.None, - useOverlayDatabaseCaching: false, - repositoryProperties: {}, - }; + }); t.deepEqual(config, expectedConfig); }); @@ -507,9 +499,7 @@ test("load non-empty input", async (t) => { }; // And the config we expect it to parse to - const expectedConfig: configUtils.Config = { - version: actionsUtil.getActionVersion(), - analysisKinds: [AnalysisKind.CodeScanning], + const expectedConfig = createTestConfig({ languages: [KnownLanguage.javascript], buildMode: BuildMode.None, originalUserInput: userConfig, @@ -521,14 +511,7 @@ test("load non-empty input", async (t) => { debugMode: false, debugArtifactName: "my-artifact", debugDatabaseName: "my-db", - trapCaches: {}, - trapCacheDownloadTime: 0, - dependencyCachingEnabled: CachingKind.None, - extraQueryExclusions: [], - overlayDatabaseMode: OverlayDatabaseMode.None, - useOverlayDatabaseCaching: false, - repositoryProperties: {}, - }; + }); const languagesInput = "javascript"; const configFilePath = createConfigFile(inputFileContents, tempDir); @@ -990,12 +973,12 @@ interface OverlayDatabaseModeTestSetup { features: Feature[]; isPullRequest: boolean; isDefaultBranch: boolean; - repositoryOwner: string; buildMode: BuildMode | undefined; languages: Language[]; codeqlVersion: string; gitRoot: string | undefined; codeScanningConfig: configUtils.UserConfig; + diskUsage: DiskUsage | undefined; } const defaultOverlayDatabaseModeTestSetup: OverlayDatabaseModeTestSetup = { @@ -1003,12 +986,15 @@ const defaultOverlayDatabaseModeTestSetup: OverlayDatabaseModeTestSetup = { features: [], isPullRequest: false, isDefaultBranch: false, - repositoryOwner: "github", buildMode: BuildMode.None, languages: [KnownLanguage.javascript], codeqlVersion: CODEQL_OVERLAY_MINIMUM_VERSION, gitRoot: "/some/git/root", codeScanningConfig: {}, + diskUsage: { + numAvailableBytes: 50_000_000_000, + numTotalBytes: 100_000_000_000, + }, }; const getOverlayDatabaseModeMacro = test.macro({ @@ -1041,6 +1027,8 @@ const getOverlayDatabaseModeMacro = test.macro({ setup.overlayDatabaseEnvVar; } + sinon.stub(util, "checkDiskUsage").resolves(setup.diskUsage); + // Mock feature flags const features = createFeatures(setup.features); @@ -1049,12 +1037,6 @@ const getOverlayDatabaseModeMacro = test.macro({ .stub(actionsUtil, "isAnalyzingPullRequest") .returns(setup.isPullRequest); - // Mock repository owner - const repository = { - owner: setup.repositoryOwner, - repo: "test-repo", - }; - // Set up CodeQL mock const codeql = mockCodeQLVersion(setup.codeqlVersion); @@ -1077,7 +1059,6 @@ const getOverlayDatabaseModeMacro = test.macro({ const result = await configUtils.getOverlayDatabaseMode( codeql, - repository, features, setup.languages, tempDir, // sourceRoot @@ -1205,6 +1186,45 @@ test( }, ); +test( + getOverlayDatabaseModeMacro, + "No overlay-base database on default branch if runner disk space is too low", + { + languages: [KnownLanguage.javascript], + features: [ + Feature.OverlayAnalysis, + Feature.OverlayAnalysisCodeScanningJavascript, + ], + isDefaultBranch: true, + diskUsage: { + numAvailableBytes: 1_000_000_000, + numTotalBytes: 100_000_000_000, + }, + }, + { + overlayDatabaseMode: OverlayDatabaseMode.None, + useOverlayDatabaseCaching: false, + }, +); + +test( + getOverlayDatabaseModeMacro, + "No overlay-base database on default branch if we can't determine runner disk space", + { + languages: [KnownLanguage.javascript], + features: [ + Feature.OverlayAnalysis, + Feature.OverlayAnalysisCodeScanningJavascript, + ], + isDefaultBranch: true, + diskUsage: undefined, + }, + { + overlayDatabaseMode: OverlayDatabaseMode.None, + useOverlayDatabaseCaching: false, + }, +); + test( getOverlayDatabaseModeMacro, "No overlay-base database on default branch when code-scanning feature enabled with disable-default-queries", @@ -1375,6 +1395,45 @@ test( }, ); +test( + getOverlayDatabaseModeMacro, + "No overlay analysis on PR if runner disk space is too low", + { + languages: [KnownLanguage.javascript], + features: [ + Feature.OverlayAnalysis, + Feature.OverlayAnalysisCodeScanningJavascript, + ], + isPullRequest: true, + diskUsage: { + numAvailableBytes: 1_000_000_000, + numTotalBytes: 100_000_000_000, + }, + }, + { + overlayDatabaseMode: OverlayDatabaseMode.None, + useOverlayDatabaseCaching: false, + }, +); + +test( + getOverlayDatabaseModeMacro, + "No overlay analysis on PR if we can't determine runner disk space", + { + languages: [KnownLanguage.javascript], + features: [ + Feature.OverlayAnalysis, + Feature.OverlayAnalysisCodeScanningJavascript, + ], + isPullRequest: true, + diskUsage: undefined, + }, + { + overlayDatabaseMode: OverlayDatabaseMode.None, + useOverlayDatabaseCaching: false, + }, +); + test( getOverlayDatabaseModeMacro, "No overlay analysis on PR when code-scanning feature enabled with disable-default-queries", @@ -1499,10 +1558,9 @@ test( test( getOverlayDatabaseModeMacro, - "Overlay PR analysis by env for dsp-testing", + "Overlay PR analysis by env", { overlayDatabaseEnvVar: "overlay", - repositoryOwner: "dsp-testing", }, { overlayDatabaseMode: OverlayDatabaseMode.Overlay, @@ -1512,10 +1570,10 @@ test( test( getOverlayDatabaseModeMacro, - "Overlay PR analysis by env for other-org", + "Overlay PR analysis by env on a runner with low disk space", { overlayDatabaseEnvVar: "overlay", - repositoryOwner: "other-org", + diskUsage: { numAvailableBytes: 0, numTotalBytes: 100_000_000_000 }, }, { overlayDatabaseMode: OverlayDatabaseMode.Overlay, @@ -1525,12 +1583,11 @@ test( test( getOverlayDatabaseModeMacro, - "Overlay PR analysis by feature flag for dsp-testing", + "Overlay PR analysis by feature flag", { languages: [KnownLanguage.javascript], features: [Feature.OverlayAnalysis, Feature.OverlayAnalysisJavascript], isPullRequest: true, - repositoryOwner: "dsp-testing", }, { overlayDatabaseMode: OverlayDatabaseMode.Overlay, @@ -1538,21 +1595,6 @@ test( }, ); -test( - getOverlayDatabaseModeMacro, - "No overlay PR analysis by feature flag for other-org", - { - languages: [KnownLanguage.javascript], - features: [Feature.OverlayAnalysis, Feature.OverlayAnalysisJavascript], - isPullRequest: true, - repositoryOwner: "other-org", - }, - { - overlayDatabaseMode: OverlayDatabaseMode.None, - useOverlayDatabaseCaching: false, - }, -); - test( getOverlayDatabaseModeMacro, "Fallback due to autobuild with traced language", diff --git a/src/config-utils.ts b/src/config-utils.ts index 0f152a633..ee9d41198 100644 --- a/src/config-utils.ts +++ b/src/config-utils.ts @@ -43,10 +43,22 @@ import { codeQlVersionAtLeast, cloneObject, isDefined, + checkDiskUsage, } from "./util"; export * from "./config/db-config"; +/** + * The minimum available disk space (in MB) required to perform overlay analysis. + * If the available disk space on the runner is below the threshold when deciding + * whether to perform overlay analysis, then the action will not perform overlay + * analysis unless overlay analysis has been explicitly enabled via environment + * variable. + */ +const OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB = 20000; +const OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES = + OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_MB * 1_000_000; + export type RegistryConfigWithCredentials = RegistryConfigNoCredentials & { // Token to use when downloading packs from this registry. token: string; @@ -148,6 +160,9 @@ export interface Config { /** A value indicating how dependency caching should be used. */ dependencyCachingEnabled: CachingKind; + /** The keys of caches that we restored, if any. */ + dependencyCachingRestoredKeys: string[]; + /** * Extra query exclusions to append to the config. */ @@ -176,7 +191,7 @@ export interface Config { repositoryProperties: RepositoryProperties; } -export async function getSupportedLanguageMap( +async function getSupportedLanguageMap( codeql: CodeQL, logger: Logger, ): Promise> { @@ -239,7 +254,7 @@ export function hasActionsWorkflows(sourceRoot: string): boolean { /** * Gets the set of languages in the current repository. */ -export async function getRawLanguagesInRepo( +async function getRawLanguagesInRepo( repository: RepositoryNwo, sourceRoot: string, logger: Logger, @@ -348,7 +363,7 @@ export function getRawLanguagesNoAutodetect( * @returns A tuple containing a list of languages in this repository that might be * analyzable and whether or not this list was determined automatically. */ -export async function getRawLanguages( +async function getRawLanguages( languagesInput: string | undefined, repository: RepositoryNwo, sourceRoot: string, @@ -496,6 +511,7 @@ export async function initActionState( trapCaches, trapCacheDownloadTime, dependencyCachingEnabled: getCachingKind(dependencyCachingEnabled), + dependencyCachingRestoredKeys: [], extraQueryExclusions: [], overlayDatabaseMode: OverlayDatabaseMode.None, useOverlayDatabaseCaching: false, @@ -579,17 +595,11 @@ const OVERLAY_ANALYSIS_CODE_SCANNING_FEATURES: Record = { }; async function isOverlayAnalysisFeatureEnabled( - repository: RepositoryNwo, features: FeatureEnablement, codeql: CodeQL, languages: Language[], codeScanningConfig: UserConfig, ): Promise { - // TODO: Remove the repository owner check once support for overlay analysis - // stabilizes, and no more backward-incompatible changes are expected. - if (!["github", "dsp-testing"].includes(repository.owner)) { - return false; - } if (!(await features.getValue(Feature.OverlayAnalysis, codeql))) { return false; } @@ -647,7 +657,6 @@ async function isOverlayAnalysisFeatureEnabled( */ export async function getOverlayDatabaseMode( codeql: CodeQL, - repository: RepositoryNwo, features: FeatureEnablement, languages: Language[], sourceRoot: string, @@ -676,27 +685,43 @@ export async function getOverlayDatabaseMode( ); } else if ( await isOverlayAnalysisFeatureEnabled( - repository, features, codeql, languages, codeScanningConfig, ) ) { - if (isAnalyzingPullRequest()) { - overlayDatabaseMode = OverlayDatabaseMode.Overlay; - useOverlayDatabaseCaching = true; + const diskUsage = await checkDiskUsage(logger); + if ( + diskUsage === undefined || + diskUsage.numAvailableBytes < OVERLAY_MINIMUM_AVAILABLE_DISK_SPACE_BYTES + ) { + const diskSpaceMb = + diskUsage === undefined + ? 0 + : Math.round(diskUsage.numAvailableBytes / 1_000_000); + overlayDatabaseMode = OverlayDatabaseMode.None; + useOverlayDatabaseCaching = false; logger.info( `Setting overlay database mode to ${overlayDatabaseMode} ` + - "with caching because we are analyzing a pull request.", - ); - } else if (await isAnalyzingDefaultBranch()) { - overlayDatabaseMode = OverlayDatabaseMode.OverlayBase; - useOverlayDatabaseCaching = true; - logger.info( - `Setting overlay database mode to ${overlayDatabaseMode} ` + - "with caching because we are analyzing the default branch.", + `due to insufficient disk space (${diskSpaceMb} MB).`, ); + } else { + if (isAnalyzingPullRequest()) { + overlayDatabaseMode = OverlayDatabaseMode.Overlay; + useOverlayDatabaseCaching = true; + logger.info( + `Setting overlay database mode to ${overlayDatabaseMode} ` + + "with caching because we are analyzing a pull request.", + ); + } else if (await isAnalyzingDefaultBranch()) { + overlayDatabaseMode = OverlayDatabaseMode.OverlayBase; + useOverlayDatabaseCaching = true; + logger.info( + `Setting overlay database mode to ${overlayDatabaseMode} ` + + "with caching because we are analyzing the default branch.", + ); + } } } @@ -846,7 +871,6 @@ export async function initConfig( const { overlayDatabaseMode, useOverlayDatabaseCaching } = await getOverlayDatabaseMode( inputs.codeql, - inputs.repository, inputs.features, config.languages, inputs.sourceRoot, @@ -1235,7 +1259,7 @@ export function isCodeQualityEnabled(config: Config): boolean { * @returns Returns `AnalysisKind.CodeScanning` if `AnalysisKind.CodeScanning` is enabled; * otherwise `AnalysisKind.CodeQuality`. */ -export function getPrimaryAnalysisKind(config: Config): AnalysisKind { +function getPrimaryAnalysisKind(config: Config): AnalysisKind { return isCodeScanningEnabled(config) ? AnalysisKind.CodeScanning : AnalysisKind.CodeQuality; diff --git a/src/database-upload.test.ts b/src/database-upload.test.ts index 6c986fb7f..e07ff1da2 100644 --- a/src/database-upload.test.ts +++ b/src/database-upload.test.ts @@ -10,11 +10,12 @@ import { GitHubApiDetails } from "./api-client"; import * as apiClient from "./api-client"; import { createStubCodeQL } from "./codeql"; import { Config } from "./config-utils"; -import { uploadDatabases } from "./database-upload"; +import { cleanupAndUploadDatabases } from "./database-upload"; import * as gitUtils from "./git-utils"; import { KnownLanguage } from "./languages"; import { RepositoryNwo } from "./repository"; import { + createFeatures, createTestConfig, getRecordingLogger, LoggedMessage, @@ -91,11 +92,12 @@ test("Abort database upload if 'upload-database' input set to false", async (t) sinon.stub(gitUtils, "isAnalyzingDefaultBranch").resolves(true); const loggedMessages = []; - await uploadDatabases( + await cleanupAndUploadDatabases( testRepoName, getCodeQL(), getTestConfig(tmpDir), testApiDetails, + createFeatures([]), getRecordingLogger(loggedMessages), ); t.assert( @@ -121,7 +123,7 @@ test("Abort database upload if 'analysis-kinds: code-scanning' is not enabled", await mockHttpRequests(201); const loggedMessages = []; - await uploadDatabases( + await cleanupAndUploadDatabases( testRepoName, getCodeQL(), { @@ -129,6 +131,7 @@ test("Abort database upload if 'analysis-kinds: code-scanning' is not enabled", analysisKinds: [AnalysisKind.CodeQuality], }, testApiDetails, + createFeatures([]), getRecordingLogger(loggedMessages), ); t.assert( @@ -155,11 +158,12 @@ test("Abort database upload if running against GHES", async (t) => { config.gitHubVersion = { type: GitHubVariant.GHES, version: "3.0" }; const loggedMessages = []; - await uploadDatabases( + await cleanupAndUploadDatabases( testRepoName, getCodeQL(), config, testApiDetails, + createFeatures([]), getRecordingLogger(loggedMessages), ); t.assert( @@ -183,11 +187,12 @@ test("Abort database upload if not analyzing default branch", async (t) => { sinon.stub(gitUtils, "isAnalyzingDefaultBranch").resolves(false); const loggedMessages = []; - await uploadDatabases( + await cleanupAndUploadDatabases( testRepoName, getCodeQL(), getTestConfig(tmpDir), testApiDetails, + createFeatures([]), getRecordingLogger(loggedMessages), ); t.assert( @@ -212,11 +217,12 @@ test("Don't crash if uploading a database fails", async (t) => { await mockHttpRequests(500); const loggedMessages = [] as LoggedMessage[]; - await uploadDatabases( + await cleanupAndUploadDatabases( testRepoName, getCodeQL(), getTestConfig(tmpDir), testApiDetails, + createFeatures([]), getRecordingLogger(loggedMessages), ); @@ -243,11 +249,12 @@ test("Successfully uploading a database to github.com", async (t) => { await mockHttpRequests(201); const loggedMessages = [] as LoggedMessage[]; - await uploadDatabases( + await cleanupAndUploadDatabases( testRepoName, getCodeQL(), getTestConfig(tmpDir), testApiDetails, + createFeatures([]), getRecordingLogger(loggedMessages), ); t.assert( @@ -272,7 +279,7 @@ test("Successfully uploading a database to GHEC-DR", async (t) => { const databaseUploadSpy = await mockHttpRequests(201); const loggedMessages = [] as LoggedMessage[]; - await uploadDatabases( + await cleanupAndUploadDatabases( testRepoName, getCodeQL(), getTestConfig(tmpDir), @@ -281,6 +288,7 @@ test("Successfully uploading a database to GHEC-DR", async (t) => { url: "https://tenant.ghe.com", apiURL: undefined, }, + createFeatures([]), getRecordingLogger(loggedMessages), ); t.assert( diff --git a/src/database-upload.ts b/src/database-upload.ts index 69175178c..d99df14c3 100644 --- a/src/database-upload.ts +++ b/src/database-upload.ts @@ -5,17 +5,20 @@ import { AnalysisKind } from "./analyses"; import { getApiClient, GitHubApiDetails } from "./api-client"; import { type CodeQL } from "./codeql"; import { Config } from "./config-utils"; +import { Feature, FeatureEnablement } from "./feature-flags"; import * as gitUtils from "./git-utils"; import { Logger, withGroupAsync } from "./logging"; +import { OverlayDatabaseMode } from "./overlay-database-utils"; import { RepositoryNwo } from "./repository"; import * as util from "./util"; -import { bundleDb, parseGitHubUrl } from "./util"; +import { bundleDb, CleanupLevel, parseGitHubUrl } from "./util"; -export async function uploadDatabases( +export async function cleanupAndUploadDatabases( repositoryNwo: RepositoryNwo, codeql: CodeQL, config: Config, apiDetails: GitHubApiDetails, + features: FeatureEnablement, logger: Logger, ): Promise { if (actionsUtil.getRequiredInput("upload-database") !== "true") { @@ -50,10 +53,16 @@ export async function uploadDatabases( return; } + const cleanupLevel = + config.overlayDatabaseMode === OverlayDatabaseMode.OverlayBase && + (await features.getValue(Feature.UploadOverlayDbToApi)) + ? CleanupLevel.Overlay + : CleanupLevel.Clear; + // Clean up the database, since intermediate results may still be written to the // database if there is high RAM pressure. await withGroupAsync("Cleaning up databases", async () => { - await codeql.databaseCleanupCluster(config, "clear"); + await codeql.databaseCleanupCluster(config, cleanupLevel); }); const client = getApiClient(); diff --git a/src/defaults.json b/src/defaults.json index 1fa392f54..835b6a33b 100644 --- a/src/defaults.json +++ b/src/defaults.json @@ -1,6 +1,6 @@ { - "bundleVersion": "codeql-bundle-v2.23.3", - "cliVersion": "2.23.3", - "priorBundleVersion": "codeql-bundle-v2.23.2", - "priorCliVersion": "2.23.2" + "bundleVersion": "codeql-bundle-v2.23.6", + "cliVersion": "2.23.6", + "priorBundleVersion": "codeql-bundle-v2.23.5", + "priorCliVersion": "2.23.5" } diff --git a/src/dependency-caching.test.ts b/src/dependency-caching.test.ts index bf2f7ba74..195cb060c 100644 --- a/src/dependency-caching.test.ts +++ b/src/dependency-caching.test.ts @@ -7,6 +7,7 @@ import test from "ava"; import * as sinon from "sinon"; import { cacheKeyHashLength } from "./caching-utils"; +import * as cachingUtils from "./caching-utils"; import { createStubCodeQL } from "./codeql"; import { CacheConfig, @@ -22,6 +23,8 @@ import { cacheKey, getCsharpDependencyDirs, getCsharpTempDependencyDir, + uploadDependencyCaches, + CacheStoreResult, } from "./dependency-caching"; import { Feature } from "./feature-flags"; import { KnownLanguage } from "./languages"; @@ -31,6 +34,7 @@ import { getRecordingLogger, checkExpectedLogMessages, LoggedMessage, + createTestConfig, } from "./testing-utils"; import { withTmpDir } from "./util"; @@ -261,15 +265,17 @@ test("downloadDependencyCaches - does not restore caches with feature keys if no .resolves(CSHARP_BASE_PATTERNS); makePatternCheckStub.withArgs(CSHARP_EXTRA_PATTERNS).resolves(undefined); - const results = await downloadDependencyCaches( + const result = await downloadDependencyCaches( codeql, createFeatures([]), [KnownLanguage.csharp], logger, ); - t.is(results.length, 1); - t.is(results[0].language, KnownLanguage.csharp); - t.is(results[0].hit_kind, CacheHitKind.Miss); + const statusReport = result.statusReport; + t.is(statusReport.length, 1); + t.is(statusReport[0].language, KnownLanguage.csharp); + t.is(statusReport[0].hit_kind, CacheHitKind.Miss); + t.deepEqual(result.restoredKeys, []); t.assert(restoreCacheStub.calledOnce); }); @@ -281,7 +287,8 @@ test("downloadDependencyCaches - restores caches with feature keys if features a const logger = getRecordingLogger(messages); const features = createFeatures([Feature.CsharpNewCacheKey]); - sinon.stub(glob, "hashFiles").resolves("abcdef"); + const mockHash = "abcdef"; + sinon.stub(glob, "hashFiles").resolves(mockHash); const keyWithFeature = await cacheKey( codeql, @@ -301,15 +308,28 @@ test("downloadDependencyCaches - restores caches with feature keys if features a .resolves(CSHARP_BASE_PATTERNS); makePatternCheckStub.withArgs(CSHARP_EXTRA_PATTERNS).resolves(undefined); - const results = await downloadDependencyCaches( + const result = await downloadDependencyCaches( codeql, features, [KnownLanguage.csharp], logger, ); - t.is(results.length, 1); - t.is(results[0].language, KnownLanguage.csharp); - t.is(results[0].hit_kind, CacheHitKind.Exact); + + // Check that the status report for telemetry indicates that one cache was restored with an exact match. + const statusReport = result.statusReport; + t.is(statusReport.length, 1); + t.is(statusReport[0].language, KnownLanguage.csharp); + t.is(statusReport[0].hit_kind, CacheHitKind.Exact); + + // Check that the restored key has been returned. + const restoredKeys = result.restoredKeys; + t.is(restoredKeys.length, 1); + t.assert( + restoredKeys[0].endsWith(mockHash), + "Expected restored key to end with hash returned by `hashFiles`", + ); + + // `restoreCache` should have been called exactly once. t.assert(restoreCacheStub.calledOnce); }); @@ -321,8 +341,14 @@ test("downloadDependencyCaches - restores caches with feature keys if features a const logger = getRecordingLogger(messages); const features = createFeatures([Feature.CsharpNewCacheKey]); + // We expect two calls to `hashFiles`: the first by the call to `cacheKey` below, + // and the second by `downloadDependencyCaches`. We use the result of the first + // call as part of the cache key that identifies a mock, existing cache. The result + // of the second call is for the primary restore key, which we don't want to match + // the first key so that we can test the restore keys logic. + const restoredHash = "abcdef"; const hashFilesStub = sinon.stub(glob, "hashFiles"); - hashFilesStub.onFirstCall().resolves("abcdef"); + hashFilesStub.onFirstCall().resolves(restoredHash); hashFilesStub.onSecondCall().resolves("123456"); const keyWithFeature = await cacheKey( @@ -343,18 +369,230 @@ test("downloadDependencyCaches - restores caches with feature keys if features a .resolves(CSHARP_BASE_PATTERNS); makePatternCheckStub.withArgs(CSHARP_EXTRA_PATTERNS).resolves(undefined); - const results = await downloadDependencyCaches( + const result = await downloadDependencyCaches( codeql, features, [KnownLanguage.csharp], logger, ); - t.is(results.length, 1); - t.is(results[0].language, KnownLanguage.csharp); - t.is(results[0].hit_kind, CacheHitKind.Partial); + + // Check that the status report for telemetry indicates that one cache was restored with a partial match. + const statusReport = result.statusReport; + t.is(statusReport.length, 1); + t.is(statusReport[0].language, KnownLanguage.csharp); + t.is(statusReport[0].hit_kind, CacheHitKind.Partial); + + // Check that the restored key has been returned. + const restoredKeys = result.restoredKeys; + t.is(restoredKeys.length, 1); + t.assert( + restoredKeys[0].endsWith(restoredHash), + "Expected restored key to end with hash returned by `hashFiles`", + ); + t.assert(restoreCacheStub.calledOnce); }); +test("uploadDependencyCaches - skips upload for a language with no cache config", async (t) => { + const codeql = createStubCodeQL({}); + const messages: LoggedMessage[] = []; + const logger = getRecordingLogger(messages); + const features = createFeatures([]); + const config = createTestConfig({ + languages: [KnownLanguage.actions], + }); + + const result = await uploadDependencyCaches(codeql, features, config, logger); + t.is(result.length, 0); + checkExpectedLogMessages(t, messages, [ + "Skipping upload of dependency cache for actions", + ]); +}); + +test("uploadDependencyCaches - skips upload if no files for the hash exist", async (t) => { + const codeql = createStubCodeQL({}); + const messages: LoggedMessage[] = []; + const logger = getRecordingLogger(messages); + const features = createFeatures([]); + const config = createTestConfig({ + languages: [KnownLanguage.go], + }); + + const makePatternCheckStub = sinon.stub(internal, "makePatternCheck"); + makePatternCheckStub.resolves(undefined); + + const result = await uploadDependencyCaches(codeql, features, config, logger); + t.is(result.length, 1); + t.is(result[0].language, KnownLanguage.go); + t.is(result[0].result, CacheStoreResult.NoHash); +}); + +test("uploadDependencyCaches - skips upload if we know the cache already exists", async (t) => { + process.env["RUNNER_OS"] = "Linux"; + + const codeql = createStubCodeQL({}); + const messages: LoggedMessage[] = []; + const logger = getRecordingLogger(messages); + const features = createFeatures([]); + + const mockHash = "abcdef"; + sinon.stub(glob, "hashFiles").resolves(mockHash); + + const makePatternCheckStub = sinon.stub(internal, "makePatternCheck"); + makePatternCheckStub + .withArgs(CSHARP_BASE_PATTERNS) + .resolves(CSHARP_BASE_PATTERNS); + + const primaryCacheKey = await cacheKey( + codeql, + features, + KnownLanguage.csharp, + CSHARP_BASE_PATTERNS, + ); + + const config = createTestConfig({ + languages: [KnownLanguage.csharp], + dependencyCachingRestoredKeys: [primaryCacheKey], + }); + + const result = await uploadDependencyCaches(codeql, features, config, logger); + t.is(result.length, 1); + t.is(result[0].language, KnownLanguage.csharp); + t.is(result[0].result, CacheStoreResult.Duplicate); +}); + +test("uploadDependencyCaches - skips upload if cache size is 0", async (t) => { + process.env["RUNNER_OS"] = "Linux"; + + const codeql = createStubCodeQL({}); + const messages: LoggedMessage[] = []; + const logger = getRecordingLogger(messages); + const features = createFeatures([]); + + const mockHash = "abcdef"; + sinon.stub(glob, "hashFiles").resolves(mockHash); + + const makePatternCheckStub = sinon.stub(internal, "makePatternCheck"); + makePatternCheckStub + .withArgs(CSHARP_BASE_PATTERNS) + .resolves(CSHARP_BASE_PATTERNS); + + sinon.stub(cachingUtils, "getTotalCacheSize").resolves(0); + + const config = createTestConfig({ + languages: [KnownLanguage.csharp], + }); + + const result = await uploadDependencyCaches(codeql, features, config, logger); + t.is(result.length, 1); + t.is(result[0].language, KnownLanguage.csharp); + t.is(result[0].result, CacheStoreResult.Empty); + + checkExpectedLogMessages(t, messages, [ + "Skipping upload of dependency cache", + ]); +}); + +test("uploadDependencyCaches - uploads caches when all requirements are met", async (t) => { + process.env["RUNNER_OS"] = "Linux"; + + const codeql = createStubCodeQL({}); + const messages: LoggedMessage[] = []; + const logger = getRecordingLogger(messages); + const features = createFeatures([]); + + const mockHash = "abcdef"; + sinon.stub(glob, "hashFiles").resolves(mockHash); + + const makePatternCheckStub = sinon.stub(internal, "makePatternCheck"); + makePatternCheckStub + .withArgs(CSHARP_BASE_PATTERNS) + .resolves(CSHARP_BASE_PATTERNS); + + sinon.stub(cachingUtils, "getTotalCacheSize").resolves(1024); + sinon.stub(actionsCache, "saveCache").resolves(); + + const config = createTestConfig({ + languages: [KnownLanguage.csharp], + }); + + const result = await uploadDependencyCaches(codeql, features, config, logger); + t.is(result.length, 1); + t.is(result[0].language, KnownLanguage.csharp); + t.is(result[0].result, CacheStoreResult.Stored); + t.is(result[0].upload_size_bytes, 1024); + + checkExpectedLogMessages(t, messages, ["Uploading cache of size"]); +}); + +test("uploadDependencyCaches - catches `ReserveCacheError` exceptions", async (t) => { + process.env["RUNNER_OS"] = "Linux"; + + const codeql = createStubCodeQL({}); + const messages: LoggedMessage[] = []; + const logger = getRecordingLogger(messages); + const features = createFeatures([]); + + const mockHash = "abcdef"; + sinon.stub(glob, "hashFiles").resolves(mockHash); + + const makePatternCheckStub = sinon.stub(internal, "makePatternCheck"); + makePatternCheckStub + .withArgs(CSHARP_BASE_PATTERNS) + .resolves(CSHARP_BASE_PATTERNS); + + sinon.stub(cachingUtils, "getTotalCacheSize").resolves(1024); + sinon + .stub(actionsCache, "saveCache") + .throws(new actionsCache.ReserveCacheError("Already in use")); + + const config = createTestConfig({ + languages: [KnownLanguage.csharp], + }); + + await t.notThrowsAsync(async () => { + const result = await uploadDependencyCaches( + codeql, + features, + config, + logger, + ); + t.is(result.length, 1); + t.is(result[0].language, KnownLanguage.csharp); + t.is(result[0].result, CacheStoreResult.Duplicate); + + checkExpectedLogMessages(t, messages, ["Not uploading cache for"]); + }); +}); + +test("uploadDependencyCaches - throws other exceptions", async (t) => { + process.env["RUNNER_OS"] = "Linux"; + + const codeql = createStubCodeQL({}); + const messages: LoggedMessage[] = []; + const logger = getRecordingLogger(messages); + const features = createFeatures([]); + + const mockHash = "abcdef"; + sinon.stub(glob, "hashFiles").resolves(mockHash); + + const makePatternCheckStub = sinon.stub(internal, "makePatternCheck"); + makePatternCheckStub + .withArgs(CSHARP_BASE_PATTERNS) + .resolves(CSHARP_BASE_PATTERNS); + + sinon.stub(cachingUtils, "getTotalCacheSize").resolves(1024); + sinon.stub(actionsCache, "saveCache").throws(); + + const config = createTestConfig({ + languages: [KnownLanguage.csharp], + }); + + await t.throwsAsync(async () => { + await uploadDependencyCaches(codeql, features, config, logger); + }); +}); + test("getFeaturePrefix - returns empty string if no features are enabled", async (t) => { const codeql = createStubCodeQL({}); const features = createFeatures([]); diff --git a/src/dependency-caching.ts b/src/dependency-caching.ts index bd39bad75..497a31fa5 100644 --- a/src/dependency-caching.ts +++ b/src/dependency-caching.ts @@ -228,6 +228,14 @@ export interface DependencyCacheRestoreStatus { /** An array of `DependencyCacheRestoreStatus` objects for each analysed language with a caching configuration. */ export type DependencyCacheRestoreStatusReport = DependencyCacheRestoreStatus[]; +/** Represents the results of `downloadDependencyCaches`. */ +export interface DownloadDependencyCachesResult { + /** The status report for telemetry */ + statusReport: DependencyCacheRestoreStatusReport; + /** An array of cache keys that we have restored and therefore know to exist. */ + restoredKeys: string[]; +} + /** * A wrapper around `cacheConfig.getHashPatterns` which logs when there are no files to calculate * a hash for the cache key from. @@ -274,8 +282,9 @@ export async function downloadDependencyCaches( features: FeatureEnablement, languages: Language[], logger: Logger, -): Promise { +): Promise { const status: DependencyCacheRestoreStatusReport = []; + const restoredKeys: string[] = []; for (const language of languages) { const cacheConfig = defaultCacheConfigs[language]; @@ -323,16 +332,27 @@ export async function downloadDependencyCaches( if (hitKey !== undefined) { logger.info(`Cache hit on key ${hitKey} for ${language}.`); - const hit_kind = - hitKey === primaryKey ? CacheHitKind.Exact : CacheHitKind.Partial; - status.push({ language, hit_kind, download_duration_ms }); + + // We have a partial cache hit, unless the key of the restored cache matches the + // primary restore key. + let hit_kind = CacheHitKind.Partial; + if (hitKey === primaryKey) { + hit_kind = CacheHitKind.Exact; + } + + status.push({ + language, + hit_kind, + download_duration_ms, + }); + restoredKeys.push(hitKey); } else { status.push({ language, hit_kind: CacheHitKind.Miss }); logger.info(`No suitable cache found for ${language}.`); } } - return status; + return { statusReport: status, restoredKeys }; } /** Enumerates possible outcomes for storing caches. */ @@ -400,6 +420,18 @@ export async function uploadDependencyCaches( continue; } + // Now that we have verified that there are suitable files, compute the hash for the cache key. + const key = await cacheKey(codeql, features, language, patterns); + + // Check that we haven't previously restored this exact key. If a cache with this key + // already exists in the Actions Cache, performing the next steps is pointless as the cache + // will not get overwritten. We can therefore skip the expensive work of measuring the size + // of the cache contents and attempting to upload it if we know that the cache already exists. + if (config.dependencyCachingRestoredKeys.includes(key)) { + status.push({ language, result: CacheStoreResult.Duplicate }); + continue; + } + // Calculate the size of the files that we would store in the cache. We use this to determine whether the // cache should be saved or not. For example, if there are no files to store, then we skip creating the // cache. In the future, we could also: @@ -425,8 +457,6 @@ export async function uploadDependencyCaches( continue; } - const key = await cacheKey(codeql, features, language, patterns); - logger.info( `Uploading cache of size ${size} for ${language} with key ${key}...`, ); diff --git a/src/environment.ts b/src/environment.ts index 16a016aaa..1d33c68a6 100644 --- a/src/environment.ts +++ b/src/environment.ts @@ -20,12 +20,6 @@ export enum EnvVar { /** Whether the CodeQL Action has invoked the Go autobuilder. */ DID_AUTOBUILD_GOLANG = "CODEQL_ACTION_DID_AUTOBUILD_GOLANG", - /** - * Whether to disable the SARIF post-processing in the Action that removes duplicate locations from - * notifications in the `run[].invocations[].toolExecutionNotifications` SARIF property. - */ - DISABLE_DUPLICATE_LOCATION_FIX = "CODEQL_ACTION_DISABLE_DUPLICATE_LOCATION_FIX", - /** * Whether the CodeQL Action is using its own deprecated and non-standard way of scanning for * multiple languages. @@ -56,20 +50,12 @@ export enum EnvVar { /** Whether the error for a deprecated version of the CodeQL Action was logged. */ LOG_VERSION_DEPRECATION = "CODEQL_ACTION_DID_LOG_VERSION_DEPRECATION", - /** - * For macOS. Result of `csrutil status` to determine whether System Integrity - * Protection is enabled. - */ - IS_SIP_ENABLED = "CODEQL_ACTION_IS_SIP_ENABLED", - /** UUID representing the current job run. */ JOB_RUN_UUID = "JOB_RUN_UUID", /** Status for the entire job, submitted to the status report in `init-post` */ JOB_STATUS = "CODEQL_ACTION_JOB_STATUS", - ODASA_TRACER_CONFIGURATION = "ODASA_TRACER_CONFIGURATION", - /** The value of the `output` input for the analyze action. */ SARIF_RESULTS_OUTPUT_DIR = "CODEQL_ACTION_SARIF_RESULTS_OUTPUT_DIR", diff --git a/src/feature-flags.ts b/src/feature-flags.ts index 27a3c0f4f..8ea1d4c1a 100644 --- a/src/feature-flags.ts +++ b/src/feature-flags.ts @@ -78,6 +78,7 @@ export enum Feature { OverlayAnalysisSwift = "overlay_analysis_swift", PythonDefaultIsToNotExtractStdlib = "python_default_is_to_not_extract_stdlib", QaTelemetryEnabled = "qa_telemetry_enabled", + UploadOverlayDbToApi = "upload_overlay_db_to_api", UseRepositoryProperties = "use_repository_properties", ValidateDbConfig = "validate_db_config", } @@ -172,6 +173,11 @@ export const featureConfig: Record< legacyApi: true, minimumVersion: undefined, }, + [Feature.JavaMinimizeDependencyJars]: { + defaultValue: false, + envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", + minimumVersion: "2.23.0", + }, [Feature.OverlayAnalysis]: { defaultValue: false, envVar: "CODEQL_ACTION_OVERLAY_ANALYSIS", @@ -283,21 +289,21 @@ export const featureConfig: Record< minimumVersion: undefined, toolsFeature: ToolsFeature.PythonDefaultIsToNotExtractStdlib, }, - [Feature.UseRepositoryProperties]: { - defaultValue: false, - envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", - minimumVersion: undefined, - }, [Feature.QaTelemetryEnabled]: { defaultValue: false, envVar: "CODEQL_ACTION_QA_TELEMETRY", legacyApi: true, minimumVersion: undefined, }, - [Feature.JavaMinimizeDependencyJars]: { + [Feature.UploadOverlayDbToApi]: { defaultValue: false, - envVar: "CODEQL_ACTION_JAVA_MINIMIZE_DEPENDENCY_JARS", - minimumVersion: "2.23.0", + envVar: "CODEQL_ACTION_UPLOAD_OVERLAY_DB_TO_API", + minimumVersion: undefined, + }, + [Feature.UseRepositoryProperties]: { + defaultValue: false, + envVar: "CODEQL_ACTION_USE_REPOSITORY_PROPERTIES", + minimumVersion: undefined, }, [Feature.ValidateDbConfig]: { defaultValue: false, diff --git a/src/git-utils.ts b/src/git-utils.ts index 837124027..0d2a7df7a 100644 --- a/src/git-utils.ts +++ b/src/git-utils.ts @@ -122,67 +122,6 @@ export const determineBaseBranchHeadCommitOid = async function ( } }; -/** - * Deepen the git history of HEAD by one level. Errors are logged. - * - * This function uses the `checkout_path` to determine the repository path and - * works only when called from `analyze` or `upload-sarif`. - */ -export const deepenGitHistory = async function () { - try { - await runGitCommand( - getOptionalInput("checkout_path"), - [ - "fetch", - "origin", - "HEAD", - "--no-tags", - "--no-recurse-submodules", - "--deepen=1", - ], - "Cannot deepen the shallow repository.", - ); - } catch { - // Errors are already logged by runGitCommand() - } -}; - -/** - * Fetch the given remote branch. Errors are logged. - * - * This function uses the `checkout_path` to determine the repository path and - * works only when called from `analyze` or `upload-sarif`. - */ -export const gitFetch = async function (branch: string, extraFlags: string[]) { - try { - await runGitCommand( - getOptionalInput("checkout_path"), - ["fetch", "--no-tags", ...extraFlags, "origin", `${branch}:${branch}`], - `Cannot fetch ${branch}.`, - ); - } catch { - // Errors are already logged by runGitCommand() - } -}; - -/** - * Repack the git repository, using with the given flags. Errors are logged. - * - * This function uses the `checkout_path` to determine the repository path and - * works only when called from `analyze` or `upload-sarif`. - */ -export const gitRepack = async function (flags: string[]) { - try { - await runGitCommand( - getOptionalInput("checkout_path"), - ["repack", ...flags], - "Cannot repack the repository.", - ); - } catch { - // Errors are already logged by runGitCommand() - } -}; - /** * Decode, if necessary, a file path produced by Git. See * https://git-scm.com/docs/git-config#Documentation/git-config.txt-corequotePath diff --git a/src/init-action.ts b/src/init-action.ts index 3512520c2..689ded2fc 100644 --- a/src/init-action.ts +++ b/src/init-action.ts @@ -371,7 +371,7 @@ async function run() { } let overlayBaseDatabaseStats: OverlayBaseDatabaseDownloadStats | undefined; - let dependencyCachingResults: DependencyCacheRestoreStatusReport | undefined; + let dependencyCachingStatus: DependencyCacheRestoreStatusReport | undefined; try { if ( config.overlayDatabaseMode === OverlayDatabaseMode.Overlay && @@ -579,12 +579,15 @@ async function run() { // Restore dependency cache(s), if they exist. if (shouldRestoreCache(config.dependencyCachingEnabled)) { - dependencyCachingResults = await downloadDependencyCaches( + const dependencyCachingResult = await downloadDependencyCaches( codeql, features, config.languages, logger, ); + dependencyCachingStatus = dependencyCachingResult.statusReport; + config.dependencyCachingRestoredKeys = + dependencyCachingResult.restoredKeys; } // Suppress warnings about disabled Python library extraction. @@ -732,7 +735,7 @@ async function run() { toolsSource, toolsVersion, overlayBaseDatabaseStats, - dependencyCachingResults, + dependencyCachingStatus, logger, error, ); @@ -755,7 +758,7 @@ async function run() { toolsSource, toolsVersion, overlayBaseDatabaseStats, - dependencyCachingResults, + dependencyCachingStatus, logger, ); } diff --git a/src/overlay-database-utils.ts b/src/overlay-database-utils.ts index ebb020ba8..a340bfe2b 100644 --- a/src/overlay-database-utils.ts +++ b/src/overlay-database-utils.ts @@ -16,6 +16,7 @@ import { type Config } from "./config-utils"; import { getCommitOid, getFileOidsUnderPath } from "./git-utils"; import { Logger, withGroupAsync } from "./logging"; import { + CleanupLevel, getErrorMessage, isInTestMode, tryGetFolderBytes, @@ -28,7 +29,7 @@ export enum OverlayDatabaseMode { None = "none", } -export const CODEQL_OVERLAY_MINIMUM_VERSION = "2.22.4"; +export const CODEQL_OVERLAY_MINIMUM_VERSION = "2.23.5"; /** * The maximum (uncompressed) size of the overlay base database that we will @@ -175,7 +176,7 @@ const MAX_CACHE_OPERATION_MS = 600_000; * @param warningPrefix Prefix for the check failure warning message * @returns True if the verification succeeded, false otherwise */ -export function checkOverlayBaseDatabase( +function checkOverlayBaseDatabase( config: Config, logger: Logger, warningPrefix: string, @@ -204,7 +205,7 @@ export function checkOverlayBaseDatabase( * @returns A promise that resolves to true if the upload was performed and * successfully completed, or false otherwise */ -export async function uploadOverlayBaseDatabaseToCache( +export async function cleanupAndUploadOverlayBaseDatabaseToCache( codeql: CodeQL, config: Config, logger: Logger, @@ -242,7 +243,7 @@ export async function uploadOverlayBaseDatabaseToCache( // Clean up the database using the overlay cleanup level. await withGroupAsync("Cleaning up databases", async () => { - await codeql.databaseCleanupCluster(config, "overlay"); + await codeql.databaseCleanupCluster(config, CleanupLevel.Overlay); }); const dbLocation = config.dbLocation; diff --git a/src/setup-codeql.ts b/src/setup-codeql.ts index 948893022..16375421a 100644 --- a/src/setup-codeql.ts +++ b/src/setup-codeql.ts @@ -34,7 +34,7 @@ export enum ToolsSource { Download = "DOWNLOAD", } -export const CODEQL_DEFAULT_ACTION_REPOSITORY = "github/codeql-action"; +const CODEQL_DEFAULT_ACTION_REPOSITORY = "github/codeql-action"; const CODEQL_NIGHTLIES_REPOSITORY_OWNER = "dsp-testing"; const CODEQL_NIGHTLIES_REPOSITORY_NAME = "codeql-cli-nightlies"; @@ -180,17 +180,6 @@ export function tryGetTagNameFromUrl( return match[1]; } -export function tryGetBundleVersionFromUrl( - url: string, - logger: Logger, -): string | undefined { - const tagName = tryGetTagNameFromUrl(url, logger); - if (tagName === undefined) { - return undefined; - } - return tryGetBundleVersionFromTagName(tagName, logger); -} - export function convertToSemVer(version: string, logger: Logger): string { if (!semver.valid(version)) { logger.debug( @@ -580,7 +569,7 @@ export async function getCodeQLSource( * Gets a fallback version number to use when looking for CodeQL in the toolcache if we didn't find * the `x.y.z` version. This is to support old versions of the toolcache. */ -export async function tryGetFallbackToolcacheVersion( +async function tryGetFallbackToolcacheVersion( cliVersion: string | undefined, tagName: string, logger: Logger, @@ -729,7 +718,7 @@ function getCanonicalToolcacheVersion( return cliVersion; } -export interface SetupCodeQLResult { +interface SetupCodeQLResult { codeqlFolder: string; toolsDownloadStatusReport?: ToolsDownloadStatusReport; toolsSource: ToolsSource; @@ -750,7 +739,7 @@ export async function setupCodeQLBundle( defaultCliVersion: CodeQLDefaultVersionInfo, features: FeatureEnablement, logger: Logger, -) { +): Promise { if (!(await util.isBinaryAccessible("tar", logger))) { throw new util.ConfigurationError( "Could not find tar in PATH, so unable to extract CodeQL bundle.", diff --git a/src/start-proxy.ts b/src/start-proxy.ts index 2888e1a58..2a082ed62 100644 --- a/src/start-proxy.ts +++ b/src/start-proxy.ts @@ -8,7 +8,7 @@ import { ConfigurationError, getErrorMessage, isDefined } from "./util"; export const UPDATEJOB_PROXY = "update-job-proxy"; export const UPDATEJOB_PROXY_VERSION = "v2.0.20250624110901"; -export const UPDATEJOB_PROXY_URL_PREFIX = +const UPDATEJOB_PROXY_URL_PREFIX = "https://github.com/github/codeql-action/releases/download/codeql-bundle-v2.22.0/"; export type Credential = { @@ -202,7 +202,7 @@ export function getFallbackUrl(proxyPackage: string): string { * * @returns The response from the GitHub API. */ -export async function getLinkedRelease() { +async function getLinkedRelease() { return getApiClient().rest.repos.getReleaseByTag({ owner: "github", repo: "codeql-action", diff --git a/src/status-report.ts b/src/status-report.ts index 1ad53ac32..c6e747489 100644 --- a/src/status-report.ts +++ b/src/status-report.ts @@ -54,7 +54,7 @@ export enum ActionName { * considered to be a third party analysis and is treated differently when calculating SLOs. To ensure * misconfigured workflows are not treated as third party, only the upload-sarif action can return false. */ -export function isFirstPartyAnalysis(actionName: ActionName): boolean { +function isFirstPartyAnalysis(actionName: ActionName): boolean { if (actionName !== ActionName.UploadSarif) { return true; } diff --git a/src/testing-utils.ts b/src/testing-utils.ts index 12193309b..eaec11e2c 100644 --- a/src/testing-utils.ts +++ b/src/testing-utils.ts @@ -392,6 +392,7 @@ export function createTestConfig(overrides: Partial): Config { trapCaches: {}, trapCacheDownloadTime: 0, dependencyCachingEnabled: CachingKind.None, + dependencyCachingRestoredKeys: [], extraQueryExclusions: [], overlayDatabaseMode: OverlayDatabaseMode.None, useOverlayDatabaseCaching: false, diff --git a/src/tools-download.ts b/src/tools-download.ts index 4cfba397e..5d8a4c5fb 100644 --- a/src/tools-download.ts +++ b/src/tools-download.ts @@ -17,7 +17,7 @@ import { cleanUpPath, getErrorMessage, getRequiredEnvParam } from "./util"; /** * High watermark to use when streaming the download and extraction of the CodeQL tools. */ -export const STREAMING_HIGH_WATERMARK_BYTES = 4 * 1024 * 1024; // 4 MiB +const STREAMING_HIGH_WATERMARK_BYTES = 4 * 1024 * 1024; // 4 MiB /** * The name of the tool cache directory for the CodeQL tools. diff --git a/src/tracer-config.ts b/src/tracer-config.ts index 9eea3eecc..d786d4651 100644 --- a/src/tracer-config.ts +++ b/src/tracer-config.ts @@ -76,7 +76,7 @@ export async function endTracingForCluster( } } -export async function getTracerConfigForCluster( +async function getTracerConfigForCluster( config: Config, ): Promise { const tracingEnvVariables = JSON.parse( diff --git a/src/upload-lib.ts b/src/upload-lib.ts index f032b8327..ac0274520 100644 --- a/src/upload-lib.ts +++ b/src/upload-lib.ts @@ -412,7 +412,7 @@ export function findSarifFilesInDir( return sarifFiles; } -export function getSarifFilePaths( +function getSarifFilePaths( sarifPath: string, isSarif: (name: string) => boolean, ) { diff --git a/src/util.test.ts b/src/util.test.ts index 03d7d89ec..2a8d941ec 100644 --- a/src/util.test.ts +++ b/src/util.test.ts @@ -476,7 +476,7 @@ for (const [ githubVersion, )}`; test(`checkActionVersion ${reportErrorDescription} for ${versionsDescription}`, async (t) => { - const warningSpy = sinon.spy(core, "error"); + const warningSpy = sinon.spy(core, "warning"); const versionStub = sinon .stub(api, "getGitHubVersion") .resolves(githubVersion); diff --git a/src/util.ts b/src/util.ts index 7136119c5..aefcc5a2a 100644 --- a/src/util.ts +++ b/src/util.ts @@ -4,7 +4,6 @@ import * as os from "os"; import * as path from "path"; import * as core from "@actions/core"; -import * as exec from "@actions/exec/lib/exec"; import * as io from "@actions/io"; import getFolderSize from "get-folder-size"; import * as yaml from "js-yaml"; @@ -1026,34 +1025,6 @@ export function fixInvalidNotifications( return newSarif; } -/** - * Removes duplicates from the sarif file. - * - * When `CODEQL_ACTION_DISABLE_DUPLICATE_LOCATION_FIX` is set to true, this will - * simply rename the input file to the output file. Otherwise, it will parse the - * input file as JSON, remove duplicate locations from the SARIF notification - * objects, and write the result to the output file. - * - * For context, see documentation of: - * `CODEQL_ACTION_DISABLE_DUPLICATE_LOCATION_FIX`. */ -export function fixInvalidNotificationsInFile( - inputPath: string, - outputPath: string, - logger: Logger, -): void { - if (process.env[EnvVar.DISABLE_DUPLICATE_LOCATION_FIX] === "true") { - logger.info( - "SARIF notification object duplicate location fix disabled by the " + - `${EnvVar.DISABLE_DUPLICATE_LOCATION_FIX} environment variable.`, - ); - fs.renameSync(inputPath, outputPath); - } else { - let sarif = JSON.parse(fs.readFileSync(inputPath, "utf8")) as SarifFile; - sarif = fixInvalidNotifications(sarif, logger); - fs.writeFileSync(outputPath, JSON.stringify(sarif)); - } -} - export function wrapError(error: unknown): Error { return error instanceof Error ? error : new Error(String(error)); } @@ -1141,7 +1112,7 @@ export function checkActionVersion( ">=3.20", )) ) { - core.error( + core.warning( "CodeQL Action v3 will be deprecated in December 2026. " + "Please update all occurrences of the CodeQL Action in your workflow files to v4. " + "For more information, see " + @@ -1197,49 +1168,6 @@ export function cloneObject(obj: T): T { return JSON.parse(JSON.stringify(obj)) as T; } -// The first time this function is called, it runs `csrutil status` to determine -// whether System Integrity Protection is enabled; and saves the result in an -// environment variable. Afterwards, simply return the value of the environment -// variable. -export async function checkSipEnablement( - logger: Logger, -): Promise { - if ( - process.env[EnvVar.IS_SIP_ENABLED] !== undefined && - ["true", "false"].includes(process.env[EnvVar.IS_SIP_ENABLED]) - ) { - return process.env[EnvVar.IS_SIP_ENABLED] === "true"; - } - - try { - const sipStatusOutput = await exec.getExecOutput("csrutil status"); - if (sipStatusOutput.exitCode === 0) { - if ( - sipStatusOutput.stdout.includes( - "System Integrity Protection status: enabled.", - ) - ) { - core.exportVariable(EnvVar.IS_SIP_ENABLED, "true"); - return true; - } - if ( - sipStatusOutput.stdout.includes( - "System Integrity Protection status: disabled.", - ) - ) { - core.exportVariable(EnvVar.IS_SIP_ENABLED, "false"); - return false; - } - } - return undefined; - } catch (e) { - logger.warning( - `Failed to determine if System Integrity Protection was enabled: ${e}`, - ); - return undefined; - } -} - export async function cleanUpPath(file: string, name: string, logger: Logger) { logger.debug(`Cleaning up ${name}.`); try { @@ -1291,17 +1219,6 @@ export function isDefined(value: T | null | undefined): value is T { return value !== undefined && value !== null; } -/** Like `Object.keys`, but typed so that the elements of the resulting array have the - * same type as the keys of the input object. Note that this may not be sound if the input - * object has been cast to `T` from a subtype of `T` and contains additional keys that - * are not represented by `keyof T`. - */ -export function unsafeKeysInvariant>( - object: T, -): Array { - return Object.keys(object) as Array; -} - /** Like `Object.entries`, but typed so that the key elements of the result have the * same type as the keys of the input object. Note that this may not be sound if the input * object has been cast to `T` from a subtype of `T` and contains additional keys that @@ -1314,3 +1231,8 @@ export function unsafeEntriesInvariant>( ([_, val]) => val !== undefined, ) as Array<[keyof T, Exclude]>; } + +export enum CleanupLevel { + Clear = "clear", + Overlay = "overlay", +} diff --git a/tests/multi-language-repo/global.json b/tests/multi-language-repo/global.json new file mode 100644 index 000000000..764fdd82b --- /dev/null +++ b/tests/multi-language-repo/global.json @@ -0,0 +1,6 @@ +{ + "sdk": { + "version": "9.0.307", + "rollForward": "latestFeature" + } +}