A concourse resource to interact with the azure blob service.
NOTE: The resource has been moved from the
czero
dockerhub account to thepcfabr
dockerhub account. If your pipeline is currently using the resource fromczero
it should be switched topcfabr
.
-
storage_account_name
: Required. The storage account name on Azure. -
storage_account_key
: Required. The storage account access key for the storage account on Azure. -
container
: Required. The name of the container in the storage account. -
base_url
: Optional. The storage endpoint to use for the resource. Defaults to the Azure Public Cloud (core.windows.net).
One of the two options must be specified:
-
regexp
: Optional. The pattern to match filenames against. At least one capture group must be specified, with parentheses to extract the version. If multiple capture groups are provided the first group is used by default, but if a group is namedversion
that will be extracted as the version. Semantic versions and numbers are supported for versioning. -
versioned_file
: Optional. The file name of the blob to be managed by the resource. The resource only pulls the latest snapshot. If the blob doesn't have a snapshot, the resource will not find the blob. A new snapshot must also be created when a blob is updated for the resource to successfully check new versions.
Checks for new versions of a file. The resource will either check snapshots when using
versioned_file
or versions in the path name when using regexp
. When using snapshots, if
a blob exists without a snapshot the resource will create a 0001-01-01T00:00:00Z
timestamp.
Places the following files in the destination:
-
(filename)
: The file fetched from the bucket. -
url
: A file containing the URL of the object. If private is true, this URL will be signed. -
version
: The version identified in the file name.
-
skip_download
: Optional. Skip downloading object. -
unpack
: Optional. If true, the blob will be unpacked before running the task. Supports tar, zip, gzip files. -
block_size
: Optional. Changes the block size used when downloading from Azure. Defaults to 4 MB. Maximum block size is 100 MB. A blob can include up to 50,000 blocks. This means with the default of 4 MB, blobs are limited to a size of a little more than 195 GB (4 MB x 50000 blocks). The max size of a blob with a block size of 100 MB will be 4.75 TB (100 MB x 50000 blocks). -
retry
:try_timeout
: Optional. Changes the try timeout in the retry options when uploading or downloading to Azure. This is the maximum allowed time for a single try of an HTTP request. A value of zero uses the default timeout. NOTE: When transferring large amounts of data, the default TryTimeout will probably not be sufficient. You should override this value based on the bandwidth available to the host machine and proximity to the Storage service. A good starting point may be something like (60 seconds per MB of anticipated-payload-size)[1]. The size of the payload would be the block size, not the overall size of the blob. This field accepts a either an integer that uses ns as the unit or a string that is a decimal number with a suffix. Valid suffixes are ns, us, ms, s, m, h.
Uploads a file to the container. If regexp
is specified, the new file will be uploaded
to the directory that the regex searches in. If versioned_file
is specified, the
new file will be uploaded as a new snapshot of that file.
-
file
: Required. Path to the file to upload, provided by an output of a task. If multiple files are matched by the glob, an error is raised. The file that matches the glob will be uploaded into the directory specified by the regexp. Only supports bash glob expansion, not regex. -
block_size
: Optional. Changes the block size used when uploading to Azure. Defaults to 4 MB. Maximum block size is 100 MB. A blob can include up to 50,000 blocks. This means with the default of 4 MB, blobs are limited to a size of a little more than 195 GB (4 MB x 50000 blocks). The max size of a blob with a block size of 100 MB will be 4.75 TB (100 MB x 50000 blocks). -
retry
:try_timeout
: Optional. Changes the try timeout in the retry options when uploading or downloading to Azure. This is the maximum allowed time for a single try of an HTTP request. A value of zero uses the default timeout. NOTE: When transferring large amounts of data, the default TryTimeout will probably not be sufficient. You should override this value based on the bandwidth available to the host machine and proximity to the Storage service. A good starting point may be something like (60 seconds per MB of anticipated-payload-size)[1]. The size of the payload would be the block size, not the overall size of the blob. This field accepts a either an integer that uses ns as the unit or a string that is a decimal number with a suffix. Valid suffixes are ns, us, ms, s, m, h.
An example pipeline exists in the example
directory.
When using Azure snapshots:
resource_types:
- name: azure-blobstore
type: docker-image
source:
repository: pcfabr/azure-blobstore-resource
resources:
- name: terraform-state
type: azure-blobstore
source:
storage_account_name: {{storage_account_name}}
storage_account_key: {{storage_account_key}}
container: {{container}}
versioned_file: terraform.tfstate
or with regexp:
resource_types:
- name: azure-blobstore
type: docker-image
source:
repository: pcfabr/azure-blobstore-resource
resources:
- name: my-release
type: azure-blobstore
source:
storage_account_name: {{storage_account_name}}
storage_account_key: {{storage_account_key}}
container: {{container}}
regexp: release-(.*).tgz
- get: terraform-state
- put: terraform-state
params:
file: terraform-state/terraform.tfstate
and with regexp:
- put: my-release
params:
file: my-release/release-*.tgz