diff --git a/.github/workflows/build-test-release.yaml b/.github/workflows/build-test-release.yaml index 9c72f4329..b4159f87d 100644 --- a/.github/workflows/build-test-release.yaml +++ b/.github/workflows/build-test-release.yaml @@ -19,7 +19,7 @@ jobs: compliance-copyrights: runs-on: ubuntu-latest steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 - uses: apache/skywalking-eyes@v0.5.0 env: GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} @@ -27,13 +27,13 @@ jobs: pre-commit: runs-on: ubuntu-latest steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 - uses: actions/setup-python@v5 with: python-version: "3.7" - run: | bash <(curl https://raw.githubusercontent.com/rhysd/actionlint/v1.6.26/scripts/download-actionlint.bash) - - uses: pre-commit/action@v3.0.0 + - uses: pre-commit/action@v3.0.1 publish: needs: @@ -41,7 +41,7 @@ jobs: - pre-commit runs-on: ubuntu-latest steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: submodules: false persist-credentials: false @@ -62,5 +62,5 @@ jobs: contents: write runs-on: ubuntu-latest steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 - uses: haya14busa/action-update-semver@v1 diff --git a/.github/workflows/reusable-build-test-release.yml b/.github/workflows/reusable-build-test-release.yml index 88e126cc5..d90f5e625 100644 --- a/.github/workflows/reusable-build-test-release.yml +++ b/.github/workflows/reusable-build-test-release.yml @@ -240,7 +240,7 @@ jobs: packages: read steps: - name: Checkout - uses: actions/checkout@v3 + uses: actions/checkout@v4 with: submodules: false persist-credentials: false @@ -257,7 +257,7 @@ jobs: GITHUB_TOKEN: ${{ github.token }} - name: Docker meta id: docker_action_meta - uses: docker/metadata-action@v5.4.0 + uses: docker/metadata-action@v5.5.1 with: images: ghcr.io/${{ github.repository }}/container tags: | @@ -283,7 +283,7 @@ jobs: - setup-workflow if: ${{ needs.setup-workflow.outputs.skip-workflow != 'Yes' }} steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 - name: run fossa analyze and create report run: | curl -H 'Cache-Control: no-cache' https://raw.githubusercontent.com/fossas/fossa-cli/master/install-latest.sh | bash @@ -292,7 +292,7 @@ jobs: env: FOSSA_API_KEY: ${{ secrets.FOSSA_API_KEY }} - name: upload THIRDPARTY file - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 with: name: THIRDPARTY path: /tmp/THIRDPARTY @@ -304,7 +304,7 @@ jobs: - fossa-scan if: ${{ needs.setup-workflow.outputs.skip-workflow != 'Yes' }} steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 - name: run fossa test run: | curl -H 'Cache-Control: no-cache' https://raw.githubusercontent.com/fossas/fossa-cli/master/install-latest.sh | bash @@ -320,7 +320,7 @@ jobs: if: ${{ needs.setup-workflow.outputs.skip-workflow != 'Yes' }} steps: - name: Checkout - uses: actions/checkout@v3 + uses: actions/checkout@v4 - name: REUSE Compliance Check uses: fsfe/reuse-action@v1.1 @@ -330,11 +330,11 @@ jobs: - setup-workflow if: ${{ needs.setup-workflow.outputs.skip-workflow != 'Yes' }} steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 - uses: actions/setup-python@v5 with: python-version: "3.7" - - uses: pre-commit/action@v3.0.0 + - uses: pre-commit/action@v3.0.1 review_secrets: name: security-detect-secrets @@ -345,13 +345,13 @@ jobs: steps: - name: Checkout if: github.event_name != 'pull_request' - uses: actions/checkout@v3 + uses: actions/checkout@v4 with: submodules: false fetch-depth: "0" - name: Checkout for PR if: github.event_name == 'pull_request' - uses: actions/checkout@v3 + uses: actions/checkout@v4 with: submodules: false fetch-depth: "0" @@ -370,7 +370,7 @@ jobs: image: returntocorp/semgrep if: ${{ needs.setup-workflow.outputs.skip-workflow != 'Yes' }} steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 - name: Semgrep id: semgrep run: semgrep ci @@ -392,7 +392,7 @@ jobs: escu: ${{ steps.testset.outputs.escu }} ucc_modinput_functional: ${{ steps.modinput-version.outputs.ucc_modinput_tests }} steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 - id: testset name: Check available test types run: | @@ -426,7 +426,7 @@ jobs: statuses: read checks: write steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} @@ -451,7 +451,7 @@ jobs: run: cp tests/unit/pytest-ci.ini pytest.ini - name: Run Pytest with coverage run: pytest --cov=./ --cov-report=xml --junitxml=test-results/junit.xml tests/unit - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: success() || failure() with: name: test-results-unit-python_${{ matrix.python-version }} @@ -476,7 +476,7 @@ jobs: statuses: read checks: write steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} @@ -501,7 +501,7 @@ jobs: run: cp tests/unit/pytest-ci.ini pytest.ini - name: Run Pytest with coverage run: pytest --cov=./ --cov-report=xml --junitxml=test-results/junit.xml tests/unit - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: success() || failure() with: name: test-results-unit-python_${{ matrix.python-version }} @@ -526,7 +526,7 @@ jobs: contents: write packages: read steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: # Very Important semantic-release won't trigger a tagged # build if this is not set false @@ -564,7 +564,7 @@ jobs: run: if [ -f requirements_dev.txt ]; then echo "ENABLED=true" >> "$GITHUB_OUTPUT"; fi - name: pip cache if: ${{ steps.checklibs.outputs.ENABLED == 'true' }} - uses: actions/cache@v3 + uses: actions/cache@v4 with: path: ${{ steps.pip-cache.outputs.dir }} key: ${{ runner.os }}-pip-${{ hashFiles('requirements_dev.txt') }} @@ -596,13 +596,13 @@ jobs: PrNumber: ${{ github.event.number }} - name: Download THIRDPARTY if: github.event_name != 'pull_request' && github.event_name != 'schedule' - uses: actions/download-artifact@v3 + uses: actions/download-artifact@v4 with: name: THIRDPARTY - name: Download THIRDPARTY (Optional for PR and schedule) if: github.event_name == 'pull_request' || github.event_name == 'schedule' continue-on-error: true - uses: actions/download-artifact@v3 + uses: actions/download-artifact@v4 with: name: THIRDPARTY - name: Update Notices @@ -620,13 +620,13 @@ jobs: source: ${{ steps.uccgen.outputs.OUTPUT }} if: always() - name: artifact-openapi - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 with: name: artifact-openapi path: ${{ github.workspace }}/${{ steps.uccgen.outputs.OUTPUT }}/appserver/static/openapi.json if: ${{ !cancelled() && needs.test-inventory.outputs.ucc_modinput_functional == 'true' && needs.test-inventory.outputs.modinput_functional == 'true' }} - name: artifact-splunk-base - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 with: name: package-splunkbase path: ${{ steps.slim.outputs.OUTPUT }} @@ -642,7 +642,7 @@ jobs: basename "${{ steps.slim.outputs.OUTPUT }}" aws s3 cp "${{ steps.slim.outputs.OUTPUT }}" s3://ta-production-artifacts/ta-apps/ - name: artifact-splunk-parts - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 with: name: package-deployment path: build/package/deployment** @@ -667,7 +667,7 @@ jobs: contents: write packages: read steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: # Very Important semantic-release won't trigger a tagged # build if this is not set false @@ -699,7 +699,7 @@ jobs: run: | echo "dir=$(pip cache dir)" >> "$GITHUB_OUTPUT" - name: pip cache - uses: actions/cache@v3 + uses: actions/cache@v4 with: path: ${{ steps.pip-cache.outputs.dir }} key: ${{ runner.os }}-pip-python3_9-${{ hashFiles('requirements_dev.txt') }} @@ -744,7 +744,7 @@ jobs: statuses: read checks: write steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 - name: Install Python 3 uses: actions/setup-python@v5 with: @@ -755,7 +755,7 @@ jobs: input-files: tests/requirement_test/logs - name: Archive production artifacts if: ${{ !cancelled() }} - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 with: name: test-results path: | @@ -779,8 +779,8 @@ jobs: - "splunk_appinspect" - "manual" steps: - - uses: actions/checkout@v3 - - uses: actions/download-artifact@v3 + - uses: actions/checkout@v4 + - uses: actions/download-artifact@v4 with: name: package-splunkbase path: build/package/ @@ -792,13 +792,13 @@ jobs: result_file: appinspect_result_${{ matrix.tags }}.json - name: upload-appinspect-report if: ${{ !cancelled() }} - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 with: name: appinspect_${{ matrix.tags }}_checks.json path: appinspect_result_${{ matrix.tags }}.json - name: upload-markdown if: matrix.tags == 'manual' - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 with: name: check_markdown path: | @@ -818,8 +818,8 @@ jobs: tags: - "cloud" steps: - - uses: actions/checkout@v3 - - uses: actions/download-artifact@v3 + - uses: actions/checkout@v4 + - uses: actions/download-artifact@v4 with: name: package-splunkbase path: build/package @@ -830,7 +830,7 @@ jobs: password: ${{ secrets.SPL_COM_PASSWORD }} app_path: build/package/ included_tags: ${{ matrix.tags }} - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: always() with: name: appinspect-api-html-report-${{ matrix.tags }} @@ -862,7 +862,7 @@ jobs: env: BUILD_NAME: ${{ needs.build.outputs.buildname }} steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: submodules: recursive token: ${{ secrets.GH_TOKEN_ADMIN }} @@ -900,7 +900,7 @@ jobs: echo "spl-host-suffix=wfe.splgdi.com" echo "k8s-manifests-branch=main" } >> "$GITHUB_OUTPUT" - - uses: actions/download-artifact@v3 + - uses: actions/download-artifact@v4 if: ${{ needs.test-inventory.outputs.ucc_modinput_functional == 'true' && needs.test-inventory.outputs.modinput_functional == 'true'}} id: download-openapi with: @@ -955,7 +955,7 @@ jobs: sc4s: ${{ fromJson(needs.meta.outputs.python39_sc4s) }} python39: true container: - image: ghcr.io/splunk/workflow-engine-base:2.0.3 + image: ghcr.io/splunk/workflow-engine-base:2.0.12 env: ARGO_SERVER: ${{ needs.setup.outputs.argo-server }} ARGO_HTTP1: ${{ needs.setup.outputs.argo-http1 }} @@ -973,7 +973,7 @@ jobs: statuses: read checks: write steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: submodules: recursive - name: capture start time @@ -981,7 +981,7 @@ jobs: run: | echo "start_time=$(date +%s)" >> "$GITHUB_OUTPUT" - name: Configure AWS credentials - uses: aws-actions/configure-aws-credentials@v2 + uses: aws-actions/configure-aws-credentials@v4 with: aws-access-key-id: ${{ secrets.AWS_ACCESS_KEY_ID }} aws-secret-access-key: ${{ secrets.AWS_SECRET_ACCESS_KEY }} @@ -1129,27 +1129,27 @@ jobs: echo "pulling logs" mkdir -p ${{ needs.setup.outputs.directory-path }}/argo-logs aws s3 cp s3://${{ needs.setup.outputs.s3-bucket }}/${WORKFLOW_NAME}/ ${{ needs.setup.outputs.directory-path }}/argo-logs/ --recursive - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ !cancelled() }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} tests artifacts path: | ${{ needs.setup.outputs.directory-path }}/test-results - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ !cancelled() }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} tests logs path: | ${{ needs.setup.outputs.directory-path }}/argo-logs - name: Upload cim-compliance-report for ${{ matrix.splunk.version }} - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 if: ${{ matrix.splunk.islatest == true }} with: name: cim-compliance-report path: | ${{ needs.setup.outputs.directory-path }}/test-results/cim-compliance-report.md - name: Upload cim-field-report for ${{ matrix.splunk.version }} - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 if: ${{ matrix.splunk.islatest == true }} with: name: cim-field-report @@ -1157,7 +1157,7 @@ jobs: ${{ needs.setup.outputs.directory-path }}/test-results/cim_field_report.json - name: Test Report id: test_report - uses: dorny/test-reporter@v1.7.0 + uses: dorny/test-reporter@v1.8.0 if: ${{ !cancelled() }} with: name: splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} test report @@ -1168,7 +1168,7 @@ jobs: run: | echo "pulling diag" aws s3 cp s3://${{ needs.setup.outputs.s3-bucket }}/diag-${{ steps.create-job-name.outputs.job-name }}/diag-${{ steps.create-job-name.outputs.job-name }}.tgz ${{ needs.setup.outputs.directory-path }}/ - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ failure() && steps.test_report.outputs.conclusion == 'failure' }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} tests diag @@ -1197,7 +1197,7 @@ jobs: sc4s: ${{ fromJson(needs.meta.outputs.python39_sc4s) }} python39: true container: - image: ghcr.io/splunk/workflow-engine-base:2.0.3 + image: ghcr.io/splunk/workflow-engine-base:2.0.12 env: ARGO_SERVER: ${{ needs.setup.outputs.argo-server }} ARGO_HTTP1: ${{ needs.setup.outputs.argo-http1 }} @@ -1214,7 +1214,7 @@ jobs: statuses: read checks: write steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: submodules: recursive - name: capture start time @@ -1222,7 +1222,7 @@ jobs: run: | echo "start_time=$(date +%s)" >> "$GITHUB_OUTPUT" - name: Configure AWS credentials - uses: aws-actions/configure-aws-credentials@v2 + uses: aws-actions/configure-aws-credentials@v4 with: aws-access-key-id: ${{ secrets.AWS_ACCESS_KEY_ID }} aws-secret-access-key: ${{ secrets.AWS_SECRET_ACCESS_KEY }} @@ -1364,13 +1364,13 @@ jobs: echo "pulling logs" mkdir -p ${{ needs.setup.outputs.directory-path }}/argo-logs aws s3 cp s3://${{ needs.setup.outputs.s3-bucket }}/${WORKFLOW_NAME}/ ${{ needs.setup.outputs.directory-path }}/argo-logs/ --recursive - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ !cancelled() }} with: name: archive splunk ${{ matrix.splunk.version }} ${{ env.TEST_TYPE }} tests artifacts path: | ${{ needs.setup.outputs.directory-path }}/test-results - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ !cancelled() }} with: name: archive splunk ${{ matrix.splunk.version }} ${{ env.TEST_TYPE }} tests logs @@ -1378,7 +1378,7 @@ jobs: ${{ needs.setup.outputs.directory-path }}/argo-logs - name: Test Report id: test_report - uses: dorny/test-reporter@v1.7.0 + uses: dorny/test-reporter@v1.8.0 if: ${{ !cancelled() }} with: name: splunk ${{ matrix.splunk.version }} ${{ env.TEST_TYPE }} test report @@ -1389,7 +1389,7 @@ jobs: run: | echo "pulling diag" aws s3 cp s3://${{ needs.setup.outputs.s3-bucket }}/diag-${{ steps.create-job-name.outputs.job-name }}/diag-${{ steps.create-job-name.outputs.job-name }}.tgz ${{ needs.setup.outputs.directory-path }}/ - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ failure() && steps.test_report.outputs.conclusion == 'failure' }} with: name: archive splunk ${{ matrix.splunk.version }} ${{ env.TEST_TYPE }} tests diag @@ -1418,7 +1418,7 @@ jobs: browser: "chrome" python39: true container: - image: ghcr.io/splunk/workflow-engine-base:2.0.3 + image: ghcr.io/splunk/workflow-engine-base:2.0.12 env: ARGO_SERVER: ${{ needs.setup.outputs.argo-server }} ARGO_HTTP1: ${{ needs.setup.outputs.argo-http1 }} @@ -1436,7 +1436,7 @@ jobs: statuses: read checks: write steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: submodules: recursive - name: capture start time @@ -1444,7 +1444,7 @@ jobs: run: | echo "start_time=$(date +%s)" >> "$GITHUB_OUTPUT" - name: Configure AWS credentials - uses: aws-actions/configure-aws-credentials@v2 + uses: aws-actions/configure-aws-credentials@v4 with: aws-access-key-id: ${{ secrets.AWS_ACCESS_KEY_ID }} aws-secret-access-key: ${{ secrets.AWS_SECRET_ACCESS_KEY }} @@ -1592,13 +1592,13 @@ jobs: echo "pulling logs" mkdir -p ${{ needs.setup.outputs.directory-path }}/argo-logs aws s3 cp s3://${{ needs.setup.outputs.s3-bucket }}/${WORKFLOW_NAME}/ ${{ needs.setup.outputs.directory-path }}/argo-logs/ --recursive - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ !cancelled() }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.browser }} ${{ matrix.vendor-version.image }} tests artifacts path: | ${{ needs.setup.outputs.directory-path }}/test-results - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ !cancelled() }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.browser }} ${{ matrix.vendor-version.image }} tests logs @@ -1606,7 +1606,7 @@ jobs: ${{ needs.setup.outputs.directory-path }}/argo-logs - name: Test Report id: test_report - uses: dorny/test-reporter@v1.7.0 + uses: dorny/test-reporter@v1.8.0 if: ${{ !cancelled() }} with: name: splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.browser }} ${{ matrix.vendor-version.image }} test report @@ -1617,7 +1617,7 @@ jobs: run: | echo "pulling diag" aws s3 cp s3://${{ needs.setup.outputs.s3-bucket }}/diag-${{ steps.create-job-name.outputs.job-name }}/diag-${{ steps.create-job-name.outputs.job-name }}.tgz ${{ needs.setup.outputs.directory-path }}/ - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ failure() && steps.test_report.outputs.conclusion == 'failure' }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.browser }} ${{ matrix.vendor-version.image }} tests diag @@ -1647,7 +1647,7 @@ jobs: modinput-type: [ "modinput_functional" ] python39: true container: - image: ghcr.io/splunk/workflow-engine-base:2.0.3 + image: ghcr.io/splunk/workflow-engine-base:2.0.12 env: ARGO_SERVER: ${{ needs.setup.outputs.argo-server }} ARGO_HTTP1: ${{ needs.setup.outputs.argo-http1 }} @@ -1665,7 +1665,7 @@ jobs: statuses: read checks: write steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: submodules: recursive - name: capture start time @@ -1673,7 +1673,7 @@ jobs: run: | echo "start_time=$(date +%s)" >> "$GITHUB_OUTPUT" - name: Configure AWS credentials - uses: aws-actions/configure-aws-credentials@v2 + uses: aws-actions/configure-aws-credentials@v4 with: aws-access-key-id: ${{ secrets.AWS_ACCESS_KEY_ID }} aws-secret-access-key: ${{ secrets.AWS_SECRET_ACCESS_KEY }} @@ -1833,13 +1833,13 @@ jobs: echo "pulling logs" mkdir -p ${{ needs.setup.outputs.directory-path }}/argo-logs aws s3 cp s3://${{ needs.setup.outputs.s3-bucket }}/${WORKFLOW_NAME}/ ${{ needs.setup.outputs.directory-path }}/argo-logs/ --recursive - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ !cancelled() }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} tests artifacts path: | ${{ needs.setup.outputs.directory-path }}/test-results - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ !cancelled() }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} tests logs @@ -1847,7 +1847,7 @@ jobs: ${{ needs.setup.outputs.directory-path }}/argo-logs - name: Test Report id: test_report - uses: dorny/test-reporter@v1.7.0 + uses: dorny/test-reporter@v1.8.0 if: ${{ !cancelled() }} with: name: splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} test report @@ -1858,7 +1858,7 @@ jobs: run: | echo "pulling diag" aws s3 cp s3://${{ needs.setup.outputs.s3-bucket }}/diag-${{ steps.create-job-name.outputs.job-name }}/diag-${{ steps.create-job-name.outputs.job-name }}.tgz ${{ needs.setup.outputs.directory-path }}/ - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ failure() && steps.test_report.outputs.conclusion == 'failure' }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} tests diag @@ -1886,7 +1886,7 @@ jobs: os: "ubuntu:22.04" python39: true container: - image: ghcr.io/splunk/workflow-engine-base:2.0.3 + image: ghcr.io/splunk/workflow-engine-base:2.0.12 env: ARGO_SERVER: ${{ needs.setup.outputs.argo-server }} ARGO_HTTP1: ${{ needs.setup.outputs.argo-http1 }} @@ -1903,7 +1903,7 @@ jobs: statuses: read checks: write steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: submodules: recursive - name: capture start time @@ -1911,7 +1911,7 @@ jobs: run: | echo "start_time=$(date +%s)" >> "$GITHUB_OUTPUT" - name: Configure AWS credentials - uses: aws-actions/configure-aws-credentials@v2 + uses: aws-actions/configure-aws-credentials@v4 with: aws-access-key-id: ${{ secrets.AWS_ACCESS_KEY_ID }} aws-secret-access-key: ${{ secrets.AWS_SECRET_ACCESS_KEY }} @@ -2068,13 +2068,13 @@ jobs: echo "pulling logs" mkdir -p ${{ needs.setup.outputs.directory-path }}/argo-logs aws s3 cp s3://${{ needs.setup.outputs.s3-bucket }}/${WORKFLOW_NAME}/ ${{ needs.setup.outputs.directory-path }}/argo-logs/ --recursive - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ !cancelled() }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} ${{ steps.os-name-version.outputs.os-name }} ${{ steps.os-name-version.outputs.os-version }} tests artifacts path: | ${{ needs.setup.outputs.directory-path }}/test-results - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ !cancelled() }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} ${{ steps.os-name-version.outputs.os-name }} ${{ steps.os-name-version.outputs.os-version }} tests logs @@ -2082,7 +2082,7 @@ jobs: ${{ needs.setup.outputs.directory-path }}/argo-logs - name: Test Report id: test_report - uses: dorny/test-reporter@v1.7.0 + uses: dorny/test-reporter@v1.8.0 if: ${{ !cancelled() }} with: name: splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} ${{ steps.os-name-version.outputs.os-name }} ${{ steps.os-name-version.outputs.os-version }} test report @@ -2093,7 +2093,7 @@ jobs: run: | echo "pulling diag" aws s3 cp s3://${{ needs.setup.outputs.s3-bucket }}/diag-${{ steps.create-job-name.outputs.job-name }}/diag-${{ steps.create-job-name.outputs.job-name }}.tgz ${{ needs.setup.outputs.directory-path }}/ - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ failure() && steps.test_report.outputs.conclusion == 'failure' }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} ${{ steps.os-name-version.outputs.os-name }} ${{ steps.os-name-version.outputs.os-version }} tests diag @@ -2115,7 +2115,7 @@ jobs: splunk: ${{ fromJson(needs.meta.outputs.matrix_supportedSplunk) }} os: [ "ubuntu:22.04", "centos:7","redhat:8.5" ] container: - image: ghcr.io/splunk/workflow-engine-base:2.0.3 + image: ghcr.io/splunk/workflow-engine-base:2.0.12 env: ARGO_SERVER: ${{ needs.setup.outputs.argo-server }} ARGO_HTTP1: ${{ needs.setup.outputs.argo-http1 }} @@ -2132,7 +2132,7 @@ jobs: statuses: read checks: write steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: submodules: recursive - name: capture start time @@ -2140,7 +2140,7 @@ jobs: run: | echo "start_time=$(date +%s)" >> "$GITHUB_OUTPUT" - name: Configure AWS credentials - uses: aws-actions/configure-aws-credentials@v2 + uses: aws-actions/configure-aws-credentials@v4 with: aws-access-key-id: ${{ secrets.AWS_ACCESS_KEY_ID }} aws-secret-access-key: ${{ secrets.AWS_SECRET_ACCESS_KEY }} @@ -2296,13 +2296,13 @@ jobs: echo "pulling logs" mkdir -p ${{ needs.setup.outputs.directory-path }}/argo-logs aws s3 cp s3://${{ needs.setup.outputs.s3-bucket }}/${WORKFLOW_NAME}/ ${{ needs.setup.outputs.directory-path }}/argo-logs/ --recursive - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ !cancelled() }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} ${{ steps.os-name-version.outputs.os-name }} ${{ steps.os-name-version.outputs.os-version }} tests artifacts path: | ${{ needs.setup.outputs.directory-path }}/test-results - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ !cancelled() }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} ${{ steps.os-name-version.outputs.os-name }} ${{ steps.os-name-version.outputs.os-version }} tests logs @@ -2310,7 +2310,7 @@ jobs: ${{ needs.setup.outputs.directory-path }}/argo-logs - name: Test Report id: test_report - uses: dorny/test-reporter@v1.7.0 + uses: dorny/test-reporter@v1.8.0 if: ${{ !cancelled() && !contains(matrix.splunk.version, 'unreleased-python3_9') }} with: name: splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} ${{ steps.os-name-version.outputs.os-name }} ${{ steps.os-name-version.outputs.os-version }} test report @@ -2319,7 +2319,7 @@ jobs: - name: Test Report Python 3.9 continue-on-error: true id: test_report_python_3_9 - uses: dorny/test-reporter@v1.7.0 + uses: dorny/test-reporter@v1.8.0 if: ${{ !cancelled() && contains(matrix.splunk.version, 'unreleased-python3_9') }} with: name: splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} ${{ steps.os-name-version.outputs.os-name }} ${{ steps.os-name-version.outputs.os-version }} test report @@ -2330,7 +2330,7 @@ jobs: run: | echo "pulling diag" aws s3 cp s3://${{ needs.setup.outputs.s3-bucket }}/diag-${{ steps.create-job-name.outputs.job-name }}/diag-${{ steps.create-job-name.outputs.job-name }}.tgz ${{ needs.setup.outputs.directory-path }}/ - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ failure() && steps.test_report.outputs.conclusion == 'failure' }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} ${{ steps.os-name-version.outputs.os-name }} ${{ steps.os-name-version.outputs.os-version }} tests diag @@ -2351,7 +2351,7 @@ jobs: matrix: splunk: ${{ fromJson(needs.meta.outputs.matrix_latestSplunk) }} container: - image: ghcr.io/splunk/workflow-engine-base:2.0.3 + image: ghcr.io/splunk/workflow-engine-base:2.0.12 env: ARGO_SERVER: ${{ needs.setup.outputs.argo-server }} ARGO_HTTP1: ${{ needs.setup.outputs.argo-http1 }} @@ -2368,7 +2368,7 @@ jobs: statuses: read checks: write steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: submodules: recursive - name: capture start time @@ -2376,7 +2376,7 @@ jobs: run: | echo "start_time=$(date +%s)" >> "$GITHUB_OUTPUT" - name: Configure AWS credentials - uses: aws-actions/configure-aws-credentials@v2 + uses: aws-actions/configure-aws-credentials@v4 with: aws-access-key-id: ${{ secrets.AWS_ACCESS_KEY_ID }} aws-secret-access-key: ${{ secrets.AWS_SECRET_ACCESS_KEY }} @@ -2536,13 +2536,13 @@ jobs: echo "pulling logs" mkdir -p ${{ needs.setup.outputs.directory-path }}/argo-logs aws s3 cp s3://${{ needs.setup.outputs.s3-bucket }}/${WORKFLOW_NAME}/ ${{ needs.setup.outputs.directory-path }}/argo-logs/ --recursive - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ steps.get-escu-detections.outputs.escu-test-run == 'true' }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} ${{ steps.os-name-version.outputs.os-name }} ${{ steps.os-name-version.outputs.os-version }} tests artifacts path: | ${{ needs.setup.outputs.directory-path }}/test-results - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ steps.get-escu-detections.outputs.escu-test-run == 'true' }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} ${{ steps.os-name-version.outputs.os-name }} ${{ steps.os-name-version.outputs.os-version }} tests logs @@ -2550,14 +2550,14 @@ jobs: ${{ needs.setup.outputs.directory-path }}/argo-logs - name: Upload results if: ${{ steps.get-escu-detections.outputs.escu-test-run == 'true' }} - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 with: name: escu-test-result path: | ${{ needs.setup.outputs.directory-path }}/test-results/escu-result.xml - name: Test Report id: test_report - uses: dorny/test-reporter@v1.7.0 + uses: dorny/test-reporter@v1.8.0 if: ${{ steps.get-escu-detections.outputs.escu-test-run == 'true' && !cancelled() && !contains(matrix.splunk.version, 'unreleased-python3_9')}} with: name: splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} test report @@ -2566,7 +2566,7 @@ jobs: - name: Test Report Python 3.9 continue-on-error: true id: test_report_python_3_9 - uses: dorny/test-reporter@v1.7.0 + uses: dorny/test-reporter@v1.8.0 if: ${{ steps.get-escu-detections.outputs.escu-test-run == 'true' && !cancelled() && contains(matrix.splunk.version, 'unreleased-python3_9')}} with: name: splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} test report @@ -2577,7 +2577,7 @@ jobs: run: | echo "pulling diag" aws s3 cp s3://${{ needs.setup.outputs.s3-bucket }}/diag-${{ steps.create-job-name.outputs.job-name }}/diag-${{ steps.create-job-name.outputs.job-name }}.tgz ${{ needs.setup.outputs.directory-path }}/ - - uses: actions/upload-artifact@v3 + - uses: actions/upload-artifact@v4 if: ${{ failure() && steps.test_report.outputs.conclusion == 'failure' }} with: name: archive splunk ${{ matrix.splunk.version }}${{ secrets.OTHER_TA_REQUIRED_CONFIGS }} ${{ env.TEST_TYPE }} ${{ matrix.vendor-version.image }} ${{ steps.os-name-version.outputs.os-name }} ${{ steps.os-name-version.outputs.os-version }} tests diag @@ -2639,7 +2639,7 @@ jobs: statuses: write steps: - name: Checkout - uses: actions/checkout@v3 + uses: actions/checkout@v4 with: submodules: false persist-credentials: false @@ -2655,14 +2655,14 @@ jobs: passphrase: ${{ secrets.SA_GPG_PASSPHRASE }} - name: Download package-deployment if: ${{ steps.semantic.outputs.new_release_published == 'true' }} - uses: actions/download-artifact@v3 + uses: actions/download-artifact@v4 id: download-package-deployment with: name: package-deployment path: download/artifacts/ - name: Download package-splunkbase if: ${{ steps.semantic.outputs.new_release_published == 'true' }} - uses: actions/download-artifact@v3 + uses: actions/download-artifact@v4 id: download-package-splunkbase with: name: package-splunkbase @@ -2671,7 +2671,7 @@ jobs: id: download-cim-compliance-report if: ${{ steps.semantic.outputs.new_release_published == 'true' }} continue-on-error: true - uses: actions/download-artifact@v3 + uses: actions/download-artifact@v4 with: name: cim-compliance-report path: download/artifacts/deployment @@ -2679,7 +2679,7 @@ jobs: id: download-cim-field-report if: ${{ steps.semantic.outputs.new_release_published == 'true' }} continue-on-error: true - uses: actions/download-artifact@v3 + uses: actions/download-artifact@v4 with: name: cim-field-report path: download/artifacts/deployment @@ -2687,7 +2687,7 @@ jobs: id: download-escu-xml if: ${{ steps.semantic.outputs.new_release_published == 'true' }} continue-on-error: true - uses: actions/download-artifact@v3 + uses: actions/download-artifact@v4 with: name: escu-test-result path: download/artifacts/deployment