Script 'mail_helper' called by obssrc Hello community, here is the log from the commit of package digger-cli for openSUSE:Factory checked in at 2025-08-28 17:19:08 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Comparing /work/SRC/openSUSE:Factory/digger-cli (Old) and /work/SRC/openSUSE:Factory/.digger-cli.new.1977 (New) ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Package is "digger-cli" Thu Aug 28 17:19:08 2025 rev:31 rq:1301661 version:0.6.121 Changes: -------- --- /work/SRC/openSUSE:Factory/digger-cli/digger-cli.changes 2025-08-25 20:40:03.977038164 +0200 +++ /work/SRC/openSUSE:Factory/.digger-cli.new.1977/digger-cli.changes 2025-08-28 17:19:41.585791857 +0200 @@ -1,0 +2,9 @@ +Thu Aug 28 05:22:27 UTC 2025 - Johannes Kastl <[email protected]> + +- Update to version 0.6.121: + * Feat/breardon2011/logging extended by @breardon2011 in #2100 + * enable debug logs by @motatoes in #2106 + * Paginate Comment Retrieval by @breardon2011 in #2107 + * S3 cache by @breardon2011 in #2111 + +------------------------------------------------------------------- Old: ---- digger-cli-0.6.120.obscpio New: ---- digger-cli-0.6.121.obscpio ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Other differences: ------------------ ++++++ digger-cli.spec ++++++ --- /var/tmp/diff_new_pack.XCCEXF/_old 2025-08-28 17:19:42.813843324 +0200 +++ /var/tmp/diff_new_pack.XCCEXF/_new 2025-08-28 17:19:42.813843324 +0200 @@ -19,7 +19,7 @@ %define executable_name digger Name: digger-cli -Version: 0.6.120 +Version: 0.6.121 Release: 0 Summary: CLI for the digger open source IaC orchestration tool License: Apache-2.0 ++++++ _service ++++++ --- /var/tmp/diff_new_pack.XCCEXF/_old 2025-08-28 17:19:42.849844833 +0200 +++ /var/tmp/diff_new_pack.XCCEXF/_new 2025-08-28 17:19:42.849844833 +0200 @@ -6,7 +6,7 @@ <param name="exclude">go.mod</param> <param name="exclude">go.work</param> <param name="exclude">go.work.sum</param> - <param name="revision">v0.6.120</param> + <param name="revision">v0.6.121</param> <param name="match-tag">v*</param> <param name="versionformat">@PARENT_TAG@</param> <param name="versionrewrite-pattern">v(.*)</param> ++++++ _servicedata ++++++ --- /var/tmp/diff_new_pack.XCCEXF/_old 2025-08-28 17:19:42.869845672 +0200 +++ /var/tmp/diff_new_pack.XCCEXF/_new 2025-08-28 17:19:42.873845839 +0200 @@ -1,7 +1,7 @@ <servicedata> <service name="tar_scm"> <param name="url">https://github.com/diggerhq/digger</param> - <param name="changesrevision">c4e270040128d102ab11e823b3131f78ef5688c3</param></service><service name="tar_scm"> + <param name="changesrevision">f1bed6685f9b6d3080f67a99dbbe2563d5035e56</param></service><service name="tar_scm"> <param name="url">https://github.com/johanneskastl/digger</param> <param name="changesrevision">8fe377068e53e2050ff4c745388d8428d2b13bb0</param></service></servicedata> (No newline at EOF) ++++++ digger-cli-0.6.120.obscpio -> digger-cli-0.6.121.obscpio ++++++ diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/digger-cli-0.6.120/action.yml new/digger-cli-0.6.121/action.yml --- old/digger-cli-0.6.120/action.yml 2025-08-22 20:33:43.000000000 +0200 +++ new/digger-cli-0.6.121/action.yml 2025-08-27 21:33:59.000000000 +0200 @@ -211,6 +211,19 @@ description: "allows overriding of the terraform cache dir which defaults to ${github.workspace}/cache" required: false default: '' + cache-dependencies-s3: + description: "Use S3 for caching terraform/terragrunt dependencies" + required: false + default: 'false' + cache-dependencies-s3-bucket: + description: "S3 bucket URL for caching (e.g., s3://mybucket/cache)" + required: false + default: '' + cache-dependencies-s3-region: + description: "AWS region for S3 cache bucket" + required: false + default: 'us-east-1' + digger-spec: description: "(orchestrator only) the spec to pass onto digger cli" @@ -316,6 +329,24 @@ digger-cache if: ${{ inputs.cache-dependencies == 'true' }} + - name: restore-s3-cache + shell: bash + run: | + BUCKET=$(echo "${{ inputs.cache-dependencies-s3-bucket }}" | sed 's|^s3://||') + REGION="${{ inputs.cache-dependencies-s3-region }}" + + SCRIPT_PATH="${{ github.action_path }}/scripts/s3-cache-download.bash" + if [ ! -f "$SCRIPT_PATH" ]; then + echo "::error::S3 cache download script not found at $SCRIPT_PATH" + echo "Please make sure the script exists and is properly installed." + exit 1 + fi + + chmod +x "$SCRIPT_PATH" + "$SCRIPT_PATH" "$BUCKET" "$REGION" "$TF_PLUGIN_CACHE_DIR" + if: ${{ inputs.cache-dependencies-s3 == 'true' }} + + # Then terraform setup happens... - name: Setup Terraform uses: hashicorp/setup-terraform@v3 with: @@ -429,8 +460,11 @@ NO_BACKEND: ${{ inputs.no-backend }} DEBUG: 'true' - TG_PROVIDER_CACHE: ${{ inputs.cache-dependencies == 'true' && 1 || 0 }} - TERRAGRUNT_PROVIDER_CACHE: ${{ inputs.cache-dependencies == 'true' && 1 || 0 }} + TG_PROVIDER_CACHE: ${{ (inputs.cache-dependencies == 'true' || inputs.cache-dependencies-s3 == 'true') && 1 || 0 }} + TERRAGRUNT_PROVIDER_CACHE: ${{ (inputs.cache-dependencies == 'true' || inputs.cache-dependencies-s3 == 'true') && 1 || 0 }} + TF_PLUGIN_CACHE_DIR: ${{ env.TF_PLUGIN_CACHE_DIR }} + TG_PROVIDER_CACHE_DIR: ${{ env.TF_PLUGIN_CACHE_DIR }} + TERRAGRUNT_PROVIDER_CACHE_DIR: ${{ env.TF_PLUGIN_CACHE_DIR }} DIGGER_RUN_SPEC: ${{inputs.digger-spec}} run: | if [[ ${{ inputs.ee }} == "true" ]]; then @@ -476,8 +510,11 @@ INPUT_DRIFT_DETECTION_SLACK_NOTIFICATION_URL: ${{ inputs.drift-detection-slack-notification-url }} INPUT_DRIFT_DETECTION_ADVANCED_SLACK_NOTIFICATION_URL: ${{ inputs.drift-detection-advanced-slack-notification-url }} NO_BACKEND: ${{ inputs.no-backend }} - TG_PROVIDER_CACHE: ${{ inputs.cache-dependencies == 'true' && 1 || 0 }} - TERRAGRUNT_PROVIDER_CACHE: ${{ inputs.cache-dependencies == 'true' && 1 || 0 }} + TG_PROVIDER_CACHE: ${{ (inputs.cache-dependencies == 'true' || inputs.cache-dependencies-s3 == 'true') && 1 || 0 }} + TERRAGRUNT_PROVIDER_CACHE: ${{ (inputs.cache-dependencies == 'true' || inputs.cache-dependencies-s3 == 'true') && 1 || 0 }} + TF_PLUGIN_CACHE_DIR: ${{ env.TF_PLUGIN_CACHE_DIR }} + TG_PROVIDER_CACHE_DIR: ${{ env.TF_PLUGIN_CACHE_DIR }} + TERRAGRUNT_PROVIDER_CACHE_DIR: ${{ env.TF_PLUGIN_CACHE_DIR }} DIGGER_RUN_SPEC: ${{inputs.digger-spec}} id: digger shell: bash @@ -524,8 +561,11 @@ INPUT_DRIFT_DETECTION_SLACK_NOTIFICATION_URL: ${{ inputs.drift-detection-slack-notification-url }} INPUT_DRIFT_DETECTION_ADVANCED_SLACK_NOTIFICATION_URL: ${{ inputs.drift-detection-advanced-slack-notification-url }} NO_BACKEND: ${{ inputs.no-backend }} - TG_PROVIDER_CACHE: ${{ inputs.cache-dependencies == 'true' && 1 || 0 }} - TERRAGRUNT_PROVIDER_CACHE: ${{ inputs.cache-dependencies == 'true' && 1 || 0 }} + TG_PROVIDER_CACHE: ${{ (inputs.cache-dependencies == 'true' || inputs.cache-dependencies-s3 == 'true') && 1 || 0 }} + TERRAGRUNT_PROVIDER_CACHE: ${{ (inputs.cache-dependencies == 'true' || inputs.cache-dependencies-s3 == 'true') && 1 || 0 }} + TF_PLUGIN_CACHE_DIR: ${{ env.TF_PLUGIN_CACHE_DIR }} + TG_PROVIDER_CACHE_DIR: ${{ env.TF_PLUGIN_CACHE_DIR }} + TERRAGRUNT_PROVIDER_CACHE_DIR: ${{ env.TF_PLUGIN_CACHE_DIR }} DIGGER_RUN_SPEC: ${{inputs.digger-spec}} id: digger-local-run shell: bash @@ -567,6 +607,23 @@ path: ${{ env.TF_PLUGIN_CACHE_DIR }} key: digger-cache-${{ hashFiles('**/cache') }} + - name: save-s3-cache + shell: bash + run: | + BUCKET=$(echo "${{ inputs.cache-dependencies-s3-bucket }}" | sed 's|^s3://||') + REGION="${{ inputs.cache-dependencies-s3-region }}" + + SCRIPT_PATH="${{ github.action_path }}/scripts/s3-cache-upload.bash" + if [ ! -f "$SCRIPT_PATH" ]; then + echo "::error::S3 cache upload script not found at $SCRIPT_PATH" + echo "Please make sure the script exists and is properly installed." + exit 1 + fi + + chmod +x "$SCRIPT_PATH" + "$SCRIPT_PATH" "$BUCKET" "$REGION" "$TF_PLUGIN_CACHE_DIR" + if: ${{ always() && inputs.cache-dependencies-s3 == 'true' }} + branding: icon: globe color: purple diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/digger-cli-0.6.120/docs/ce/howto/caching-strategies.mdx new/digger-cli-0.6.121/docs/ce/howto/caching-strategies.mdx --- old/digger-cli-0.6.120/docs/ce/howto/caching-strategies.mdx 2025-08-22 20:33:43.000000000 +0200 +++ new/digger-cli-0.6.121/docs/ce/howto/caching-strategies.mdx 2025-08-27 21:33:59.000000000 +0200 @@ -21,6 +21,17 @@ cache is immutable so new caches will be new entries in the cache. Digger currently does not prune older caches so you may need to clean up older caches using a certain cycle. +## S3 caching + +You can cache in AWS s3 rather than artefacts. This requires a bucket for caching and setting similar settings: + +``` - uses: diggerhq/digger@vLatest + with: + cache-dependencies-s3: true + cache-dependencies-s3-bucket: s3://terraform-cache-1756322349465/cache + cache-dependencies-s3-region: us-east-1 +``` + ## Self-hosted runners and volumes caching If you are using self-hosted runners with github it might be overkill on your bandwidth to load and restore from cache. In this case it might be better to mount diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/digger-cli-0.6.120/scripts/s3-cache-download.bash new/digger-cli-0.6.121/scripts/s3-cache-download.bash --- old/digger-cli-0.6.120/scripts/s3-cache-download.bash 1970-01-01 01:00:00.000000000 +0100 +++ new/digger-cli-0.6.121/scripts/s3-cache-download.bash 2025-08-27 21:33:59.000000000 +0200 @@ -0,0 +1,45 @@ +#!/bin/bash +set -e + +# S3 Cache Download Script +# Downloads terraform/terragrunt provider cache from S3 + +BUCKET="$1" +REGION="$2" +CACHE_DIR="$3" + +# Validation +if [ -z "$BUCKET" ]; then + echo "Error: S3 bucket name is required" + exit 1 +fi + +if [ -z "$REGION" ]; then + echo "Error: AWS region is required" + exit 1 +fi + +if [ -z "$CACHE_DIR" ]; then + echo "Error: Cache directory path is required" + exit 1 +fi + +# Ensure cache directory exists +mkdir -p "$CACHE_DIR" +echo "Ensuring cache directory exists: $CACHE_DIR" + +# Download cache from S3 +echo "Restoring cache from S3 bucket: $BUCKET (region: $REGION)" +if aws s3 sync "s3://$BUCKET" "$CACHE_DIR" --region "$REGION" --quiet 2>/dev/null; then + CACHED_FILES=$(find "$CACHE_DIR" -type f 2>/dev/null | wc -l) + echo "Cache restored successfully ($CACHED_FILES files)" + + if [ "$CACHED_FILES" -gt 0 ]; then + echo "Sample cached artifacts:" + find "$CACHE_DIR" -type f 2>/dev/null | head -3 + fi +else + echo "No existing cache found or failed to restore (this is normal for first run)" +fi + +echo "Cache download completed" diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' '--exclude=.svnignore' old/digger-cli-0.6.120/scripts/s3-cache-upload.bash new/digger-cli-0.6.121/scripts/s3-cache-upload.bash --- old/digger-cli-0.6.120/scripts/s3-cache-upload.bash 1970-01-01 01:00:00.000000000 +0100 +++ new/digger-cli-0.6.121/scripts/s3-cache-upload.bash 2025-08-27 21:33:59.000000000 +0200 @@ -0,0 +1,66 @@ +#!/bin/bash +set -e + +# S3 Cache Upload Script +# Uploads terraform/terragrunt provider cache to S3 + +BUCKET="$1" +REGION="$2" +CACHE_DIR="$3" + +# Validation +if [ -z "$BUCKET" ]; then + echo "Error: S3 bucket name is required" + exit 1 +fi + +if [ -z "$REGION" ]; then + echo "Error: AWS region is required" + exit 1 +fi + +if [ -z "$CACHE_DIR" ]; then + echo "Error: Cache directory path is required" + exit 1 +fi + +# Check if cache directory exists and has files +if [ ! -d "$CACHE_DIR" ]; then + echo "No cache directory found at: $CACHE_DIR" + echo "Nothing to upload" + exit 0 +fi + +ARTIFACT_COUNT=$(find "$CACHE_DIR" -type f 2>/dev/null | wc -l) + +if [ "$ARTIFACT_COUNT" -eq 0 ]; then + echo "No files to cache - cache directory is empty" + echo "Nothing to upload" + exit 0 +fi + +# Check if AWS CLI is available and credentials are configured +if ! command -v aws &> /dev/null; then + echo "Error: AWS CLI is not installed or not in PATH" + exit 1 +fi + +# Verify AWS credentials are available +if ! aws sts get-caller-identity --region "$REGION" &> /dev/null; then + echo "Error: AWS credentials are not properly configured or are invalid" + echo "Please ensure AWS credentials are set up correctly before running this script" + exit 1 +fi + +# Upload cache to S3 +echo "Saving cache to S3 bucket: $BUCKET (region: $REGION)" +echo "Uploading $ARTIFACT_COUNT files" + +if aws s3 sync "$CACHE_DIR" "s3://$BUCKET" --region "$REGION" --quiet 2>/dev/null; then + echo "Cache saved successfully" +else + echo "Warning: Failed to save cache (this won't fail the build)" + exit 0 # Don't fail the build on cache upload failure +fi + +echo "Cache upload completed" ++++++ digger-cli.obsinfo ++++++ --- /var/tmp/diff_new_pack.XCCEXF/_old 2025-08-28 17:19:43.889888421 +0200 +++ /var/tmp/diff_new_pack.XCCEXF/_new 2025-08-28 17:19:43.893888589 +0200 @@ -1,5 +1,5 @@ name: digger-cli -version: 0.6.120 -mtime: 1755887623 -commit: c4e270040128d102ab11e823b3131f78ef5688c3 +version: 0.6.121 +mtime: 1756323239 +commit: f1bed6685f9b6d3080f67a99dbbe2563d5035e56 ++++++ vendor.tar.gz ++++++ /work/SRC/openSUSE:Factory/digger-cli/vendor.tar.gz /work/SRC/openSUSE:Factory/.digger-cli.new.1977/vendor.tar.gz differ: char 132, line 1
