azure-blob-to-s3
Batch copy files from Azure Blob Storage to Amazon S3
- Fully streaming
- Lists files from Azure Blob storage only as needed
- Uploads Azure binary data to S3 streamingly
- Skips unnecessary uploads (files with a matching key and
Content-Length
already on S3) - Retries on (frequent) failed downloads from Azure
- Generates ndjson logs for each network operation
Install
$ npm install --save azure-blob-to-s3
Usage
API
var toS3 = require('azure-blob-to-s3')
toS3({
azure: {
connection: '',
container: 'my-container'
},
aws: {
region: 'us-west-2',
bucket: 'my-bucket'
}
})
CLI
azure-s3 \
--concurrency 10 \
--azure-connection "..." \
--azure-container my-container \
--aws-bucket my-bucket
API
toS3(options)
-> output
Copys files from Azure Blob Storage to AWS S3.
options
Required
Type: object
Options for configuring the copy.
concurrency
Type: number
Default: 100
The maximum number of files to concurrently stream from Azure and into S3. This is the highWaterMark
of the file upload stream.
azure
Type: object
connection
Required
Type: string
Azure Blob Storage connection string.
container
Required
Type: string
Azure Blob Storage container name.
token
Type: string
Default: undefined
A page token where the file list operation will begin.
aws
Type: object
License
MIT © Ben Drucker