See TESTING.md.
Dpl supports the following providers:
- Anynines
- AppFog
- Atlas by HashiCorp
- AWS CodeDeploy
- AWS Elastic Beanstalk
- AWS OpsWorks
- Azure Web Apps
- Bintray
- BitBalloon
- Boxfuse
- Catalyze
- Chef Supermarket
- Cloud 66
- Cloud Foundry
- Deis
- Divshot.io
- Engine Yard
- ExoScale
- Firebase
- Github Releases
- Google App Engine (experimental)
- Google Cloud Storage
- Hackage
- Heroku
- Lambda
- Launchpad
- Modulus
- Nodejitsu
- NPM
- OpenShift
- packagecloud
- Puppet Forge
- PyPi
- Rackspace Cloud Files
- RubyGems
- S3
- Scalingo
- Script
- Surge.sh
- TestFairy
Dpl is published to rubygems.
- Dpl requires ruby with a version greater than 1.9.3
- To install:
gem install dpl
###Security Warning:
Running dpl in a terminal that saves history is insecure as your password/api key will be saved as plain text by it.
###Global Flags
--provider=<provider>
sets the provider you want to deploy to. Every provider has slightly different flags, which are documented in the section about your provider following.- Dpl will deploy by default from the latest commit. Use the
--skip_cleanup
flag to deploy from the current file state. Note that many providers deploy by git and could ignore this option.
- api-key: Heroku API Key
- strategy[git/anvil]: Deployment strategy for Dpl. Defaults to anvil.
- app: Heroku app name. Defaults to the name of your git repo.
- username: heroku username. Not necessary if api-key is used. Requires git strategy.
- password: heroku password. Not necessary if api-key is used. Requires git strategy.
- Anvil will run the buildpack compilation step on the Travis CI VM, whereas the Git strategy will run it on a Heroku dyno, which provides the same environment the application will then run under and might be slightly faster.
- The Git strategy allows using user and password instead of api-key.
- When using Git, Heroku might send you an email for every deploy, as it adds a temporary SSH key to your account.
As a rule of thumb, you should switch to the Git strategy if you run into issues with Anvil or if you're using the user-env-compile plugin.
dpl --provider=heroku --api-key=`heroku auth:token`
dpl --provider=heroku --strategy=git --username=<username> --password=<password> --app=<application>
- file: Path to a descriptor file, containing information for the Bintray upload.
- user: Bintray user
- key: Bintray API key
- passphrase: Optional. In case a passphrase is configured on Bintray and GPG signing is used.
- dry-run: Optional. If set to true, skips sending requests to Bintray. Useful for testing your configuration.
{
/* Bintray package information.
In case the package already exists on Bintray, only the name, repo and subject
fields are mandatory. */
"package": {
"name": "auto-upload", // Bintray package name
"repo": "myRepo", // Bintray repository name
"subject": "myBintrayUser", // Bintray subject (user or organization)
"desc": "I was pushed completely automatically",
"website_url": "www.jfrog.com",
"issue_tracker_url": "https://github.com/bintray/bintray-client-java/issues",
"vcs_url": "https://github.com/bintray/bintray-client-java.git",
"github_use_tag_release_notes": true,
"github_release_notes_file": "RELEASE.txt",
"licenses": ["MIT"],
"labels": ["cool", "awesome", "gorilla"],
"public_download_numbers": false,
"public_stats": false,
"attributes": [{"name": "att1", "values" : ["val1"], "type": "string"},
{"name": "att2", "values" : [1, 2.2, 4], "type": "number"},
{"name": "att5", "values" : ["2014-12-28T19:43:37+0100"], "type": "date"}]
},
/* Package version information.
In case the version already exists on Bintray, only the name fields is mandatory. */
"version": {
"name": "0.5",
"desc": "This is a version",
"released": "2015-01-04",
"vcs_tag": "0.5",
"attributes": [{"name": "VerAtt1", "values" : ["VerVal1"], "type": "string"},
{"name": "VerAtt2", "values" : [1, 3.2, 5], "type": "number"},
{"name": "VerAtt3", "values" : ["2015-01-01T19:43:37+0100"], "type": "date"}],
"gpgSign": false
},
/* Configure the files you would like to upload to Bintray and their upload path.
You can define one or more groups of patterns.
Each group contains three patterns:
includePattern: Pattern in the form of Ruby regular expression, indicating the path of files to be uploaded to Bintray.
excludePattern: Optional. Pattern in the form of Ruby regular expression, indicating the path of files to be removed from the list of files specified by the includePattern.
uploadPattern: Upload path on Bintray. The path can contain symbols in the form of $1, $2,... that are replaced with capturing groups defined in the include pattern.
In the example below, the following files are uploaded,
1. All gem files located under build/bin/ (including sub directories),
except for files under a the do-not-deploy directory.
The files will be uploaded to Bintray under the gems folder.
2. All files under build/docs. The files will be uploaded to Bintray under the docs folder.
Note: Regular expressions defined as part of the includePattern property must be wrapped with brackets. */
"files":
[
{"includePattern": "build/bin(.*)*/(.*\.gem)", "excludePattern": ".*/do-not-deploy/.*", "uploadPattern": "gems/$2"},
{"includePattern": "build/docs/(.*)", "uploadPattern": "docs/$1"}
],
"publish": true
}
When artifacts are uploaded to a Debian repository using the Automatic index layout, the Debian distribution information is required and must be specified. The information is specified in the descriptor file by the matrixParams as part of the files closure as shown in the following example:
"files":
[{"includePattern": "build/bin/(.*\.deb)", "uploadPattern": "$1",
"matrixParams": {
"deb_distribution": "vivid",
"deb_component": "main",
"deb_architecture": "amd64"}
}
]
dpl --provider=bintray --file=<path> --user=<username> --key=<api-key>
dpl --provider=bintray --file=<path> --user=<username> --key=<api-key> --passphrase=<passphrase>
Boxfuse will transform your .jar or .war file of your JVM-based application into a minimal machine image based upon which it will launch EC2 instances on AWS.
- user: Your Boxfuse user
- secret: Your Boxfuse secret
- configfile: The Boxfuse configuration file to use (default: boxfuse.conf)
- payload: The file to use as a payload for the image
- app: The Boxfuse app to deploy (default: auto-detected based on payload file name)
- version: The version to assign to the image (default: auto-detected based on payload file name)
- env: The Boxfuse environment to deploy to (default: test)
All options can also be configured directly in boxfuse.conf as described in the documentation.
For authentication you can also use Travis CI secure environment variable:
- BOXFUSE_USER: Your Boxfuse user
- BOXFUSE_SECRET: Your Boxfuse secret
dpl --provider=boxfuse
dpl --provider=boxfuse --user=<your-boxfuse-user> --secret=<your-boxfuse-secret> --env=<your-boxfuse-environment>
dpl --provider=boxfuse --configfile=<your-boxfuse-config-file>
- username: Nodejitsu Username
- api-key: Nodejitsu API Key
dpl --provider=nodejitsu --username=<username> --api-key=<api-key>
- api-key Modulus Authentication Token
- project-name Modulus Project to Deploy
dpl --provider=modulus --api-key=<api-key> --project-name=<project-name>
- api-key: Engine Yard Api Key
- username: Engine Yard username. Not necessary if api-key is used. Requires git strategy.
- password: Engine Yard password. Not necessary if api-key is used.
- app: Engine Yard Application name. Defaults to git repo's name.
- environment: Engine Yard Application Environment. Optional.
- migrate: Engine Yard migration commands. Optional.
dpl --provider=engineyard --api-key=<api-key>
dpl --provider=engineyard --username=<username> --password=<password> --environment=staging
dpl --provider=engineyard --api-key=<api-key> --app=<application> --migrate=`rake db:migrate`
- user: Openshift Username.
- password: Openshift Password.
- domain: Openshift Application Domain.
- app: Openshift Application. Defaults to git repo's name.
####Examples:
dpl --provider=openshift --user=<username> --password=<password> --domain=<domain>
dpl --provider=openshift --user=<username> --password=<password> --domain=<domain> --app=<app>
- api-key: Rubygems Api Key.
dpl --provider=rubygems --api-key=<api-key>
- user: PyPI Username.
- password: PyPI Password.
- server: Optional. Only required if you want to release to a different index. Follows the form of 'https://mypackageindex.com/index'. Defaults to 'https://pypi.python.org/pypi'.
- distributions: Optional. A space-separated list of distributions to be uploaded to PyPI. Defaults to 'sdist'.
- docs_dir: Optional. A path to the directory to upload documentation from. Defaults to 'build/docs'
- PYPI_USER: PyPI Username. Used if the
user
option is omitted. - PYPI_PASSWORD: PyPI Password. Used if the
password
option is omitted. - PYPI_SERVER Optional. Only required if you want to release to a different index. Used if the
server
option is omitted. - PYPI_DISTRIBUTIONS Optional. A space-separated list of distributions to be uploaded to PyPI. Used if the
distributions
option is omitted. - PYPI_DOCS_DIR Optional. A path to the directory to upload documentation from. Used if the
docs_dir
option is omitted.
dpl --provider=pypi --user=<username> --password=<password>
dpl --provider=pypi --user=<username> --password=<password> --server='https://mypackageindex.com/index' --distributions='sdist bdist_wheel'
- email: NPM email.
- api-key: NPM api key. Can be retrieved from your ~/.npmrc file.
dpl --provider=npm --email=<email> --api-key=<api-key>
- access-key-id: AWS Access Key ID. Can be obtained from here.
- secret-access-key: AWS Secret Key. Can be obtained from here.
- bucket: S3 Bucket.
- region: S3 Region. Defaults to us-east-1.
- upload-dir: S3 directory to upload to. Defaults to root directory.
- storage-class: S3 storage class to upload as. Defaults to "STANDARD". Other values are "STANDARD_IA" or "REDUCED_REDUNDANCY". Details can be found here.
- local-dir: Local directory to upload from. Can be set from a global perspective (~/travis/build) or relative perspective (build) Defaults to project root.
- detect-encoding: Set HTTP header
Content-Encoding
for files compressed withgzip
andcompress
utilities. Defaults to not set. - cache_control: Set HTTP header
Cache-Control
to suggest that the browser cache the file. Defaults tono-cache
. Valid options areno-cache
,no-store
,max-age=<seconds>
,s-maxage=<seconds>
no-transform
,public
,private
. - expires: This sets the date and time that the cached object is no longer cacheable. Defaults to not set. The date must be in the format
YYYY-MM-DD HH:MM:SS -ZONE
. - acl: Sets the access control for the uploaded objects. Defaults to
private
. Valid options areprivate
,public_read
,public_read_write
,authenticated_read
,bucket_owner_read
,bucket_owner_full_control
. - dot_match: When set to
true
, upload files starting a.
. - index_document_suffix: Set the index document of a S3 website.
- default_text_charset: Set the default character set to append to the content-type of text files you are uploading.
It is possible to set file-specific Cache-Control
and Expires
headers using value: file[, file]
format.
- AWS_ACCESS_KEY_ID: AWS Access Key ID. Used if the
access-key-id
option is omitted. - AWS_SECRET_ACCESS_KEY: AWS Secret Key. Used if the
secret-access-key
option is omitted.
--cache_control="no-cache: index.html"
--expires="\"2012-12-21 00:00:00 -0000\": *.css, *.js"
dpl --provider=s3 --access-key-id=<access-key-id> --secret-access-key=<secret-access-key> --bucket=<bucket> --acl=public_read
dpl --provider=s3 --access-key-id=<access-key-id> --secret-access-key=<secret-access-key> --bucket=<bucket> --detect-encoding --cache_control=max-age=99999 --expires="2012-12-21 00:00:00 -0000"
dpl --provider=s3 --access-key-id=<access-key-id> --secret-access-key=<secret-access-key> --bucket=<bucket> --region=us-west-2 --local-dir= BUILD --upload-dir=BUILDS
- access-key-id: AWS Access Key ID. Can be obtained from here.
- secret-access-key: AWS Secret Key. Can be obtained from here.
- region: AWS Region the Elastic Beanstalk app is running in. Defaults to 'us-east-1'. Please be aware that this must match the region of the elastic beanstalk app.
- app: Elastic Beanstalk application name.
- env: Elastic Beanstalk environment name which will be updated.
- zip_file: The zip file that you want to deploy. Note: you also need to use the
skip_cleanup
or the zip file you are trying to upload will be removed during cleanup. - bucket_name: Bucket name to upload app to.
- bucket_path: Location within Bucket to upload app to.
- only_create_app_version: only create the app version, don't actually deploy it.
- ELASTIC_BEANSTALK_ENV: Elastic Beanstalk environment name which will be updated. Is only used if
env
option is omitted. - ELASTIC_BEANSTALK_LABEL: Label name of the new version.
- ELASTIC_BEANSTALK_DESCRIPTION: Description of the new version. Defaults to the last commit message.
dpl --provider=elasticbeanstalk --access-key-id=<access-key-id> --secret-access-key="<secret-access-key>" --app="example-app-name" --env="example-app-environment" --region="us-west-2"
- access-key-id: AWS Access Key ID. Can be obtained from here.
- secret-access-key: AWS Secret Key. Can be obtained from here.
- app-id: The app ID.
- instance-ids: An instance id. (Use this option multiple times to specify multiple instance ids. Default: [])
- layer-ids: A layer id. (Use this option multiple times to specify multiple layer ids. Default: [])
- migrate: Migrate the database. (Default: false)
- wait-until-deployed: Wait until the app is deployed and return the deployment status. (Default: false)
- custom_json: Override custom_json options. If using this, default configuration will be overriden. See the code here. More about
custom_json
here.
- AWS_ACCESS_KEY_ID: AWS Access Key ID. Used if the
access-key-id
option is omitted. - AWS_SECRET_ACCESS_KEY: AWS Secret Key. Used if the
secret-access-key
option is omitted.
dpl --provider=opsworks --access-key-id=<access-key-id> --secret-access-key=<secret-access-key> --app-id=<app-id> --migrate --wait-until-deployed
dpl --provider=opsworks --access-key-id=<access-key-id> --secret-access-key=<secret-access-key> --app-id=<app-id> --layer-ids=<layer-id>
- username: anynines username.
- password: anynines password.
- organization: anynines target organization.
- space: anynines target space
dpl --provider=anynines --username=<username> --password=<password> --organization=<organization> --space=<space>
- email: Appfog Email.
- password: Appfog Password.
- app: Appfog App. Defaults to git repo's name.
dpl --provider=appfog --email=<email> --password=<password>
dpl --provider=appfog --email=<email> --password=<password> --app=<app>
The Atlas provider uses the atlas-upload-cli
command. The Atlas Upload CLI is a lightweight command line interface for uploading application code to Atlas to kick off Atlas-based deployment processes from Travis CI.
You first need to create an Atlas account, then, generate an Atlas API token for Travis CI.
- token (Required): Atlas API token.
- app (Required): Atlas application name (
<atlas-username>/<app-name>
). - exclude: Glob pattern of files or directories to exclude (this may be specified multiple times).
- include: Glob pattern of files/directories to include (this may be specified multiple times, any excludes will override conflicting includes).
- address: The address of the Atlas server.
- vcs: Use VCS to determine which files to include/exclude.
- metadata: Arbitrary key-value (string) metadata to be sent with the upload; may be specified multiple times.
- debug: Turn on debug output.
- version: Print the version of this application.
dpl --provider=atlas --token=ATLAS_TOKEN --app=ATLAS_USERNAME/APP_NAME
dpl --provider=atlas --token=ATLAS_TOKEN --app=ATLAS_USERNAME/APP_NAME --debug --vcs --version
dpl --provider=atlas --token=ATLAS_TOKEN --app=ATLAS_USERNAME/APP_NAME --exclude="*.log" --include="build/*" --include="bin/*"
dpl --provider=atlas --token=ATLAS_TOKEN --app=ATLAS_USERNAME/APP_NAME --metadata="foo=bar" --metadata="bar=baz"
- site: Web App Name (if your app lives at myapp.azurewebsites.net, the name would be myapp).
- slot: Optional. Slot name if your app uses staging deployment. (e.g. if your slot lives at myapp-test.azurewebsites.net, the slot would be myapp-test).
- username: Web App Deployment Username.
- password: Web App Deployment Password.
- verbose: If passed, Azure's deployment output will be printed. Warning: If you provide incorrect credentials, Git will print those in clear text. Correct authentication credentials will remain hidden.
- AZURE_WA_SITE Web App Name. Used if the
site
option is omitted. - AZURE_WA_SLOT Optional. Slot name if your app uses staging deployment. Used if the
slot
option is omitted. - AZURE_WA_USERNAME: Web App Deployment Username. Used if the
username
option is omitted. - AZURE_WA_PASSWORD: Web App Deployment Password. Used if the
password
option is omitted.
dpl --provider=AzureWebApps --username=depluser --password=deplp@ss --site=dplsite --slot=dplsite-test --verbose
- api-key: Divshot.io API key
- environment: Which environment (development, staging, production) to deploy to
dpl --provider=divshot --api-key=<api-key> --environment=<environment>
- username: Cloud Foundry username.
- password: Cloud Foundry password.
- organization: Cloud Foundry target organization.
- api: Cloud Foundry api URL
- space: Cloud Foundry target space
- manifest: Path to manifest file. Optional.
- skip_ssl_validation: Skip ssl validation. Optional.
dpl --provider=cloudfoundry --username=<username> --password=<password> --organization=<organization> --api=<api> --space=<space> --skip-ssl-validation
- username: Rackspace Username.
- api-key: Rackspace API Key.
- region: Cloud Files Region. The region in which your Cloud Files container exists.
- container: Container Name. The container where you would like your files to be uploaded.
- dot_match: When set to
true
, upload files starting a.
.
dpl --provider=cloudfiles --username=<username> --api-key=<api-key> --region=<region> --container=<container>
- api-key: GitHub oauth token with
public_repo
orrepo
permission. - user: GitHub username. Not necessary if
api-key
is used. - password: GitHub Password. Not necessary if
api-key
is used. - repo: GitHub Repo. Defaults to git repo's name.
- file: File to upload to GitHub Release.
- file_glob: If files should be interpreted as globs (* and ** wildcards). Defaults to false.
- overwrite: If files with the same name should be overwritten. Defaults to false.
- release-number: Overide automatic release detection, set a release manually.
Additionally, options can be passed to Octokit client. These are documented in https://github.com/octokit/octokit.rb/blob/master/lib/octokit/client/releases.rb.
For accounts using two factor authentication, you have to use an oauth token as a username and password will not work.
dpl --provider=releases --api-key=<api-key> --file=build.tar.gz
- redeployment_hook: The redeployment hook URL. Available from the Information menu within the Cloud 66 portal.
dpl --provider=cloud66 --redeployment_hook=<url>
- username: Hackage username.
- password: Hackage password.
dpl --provider=hackage --username=<username> --password=<password>
- controller: Deis controller e.g. deis.deisapps.com
- username: Deis username
- password: Deis password
- app: Deis app
- cli_version: Install a specific deis cli version
dpl --provider=deis --controller=deis.deisapps.com --username=travis --password=secret --app=example
- access-key-id: GCS Interoperable Access Key ID. Info about Interoperable Access Key from here.
- secret-access-key: GCS Interoperable Access Secret.
- bucket: GCS Bucket.
- upload-dir: GCS directory to upload to. Defaults to root directory.
- local-dir: Local directory to upload from. Can be set from a global perspective (~/travis/build) or relative perspective (build) Defaults to project root.
- dot_match: When set to
true
, upload files starting a.
. - detect-encoding: Set HTTP header
Content-Encoding
for files compressed withgzip
andcompress
utilities. Defaults to not set. - cache_control: Set HTTP header
Cache-Control
to suggest that the browser cache the file. Defaults to not set. Info is here - acl: Sets the access control for the uploaded objects. Defaults to not set. Info is here
dpl --provider=gcs --access-key-id=<access-key-id> --secret-access-key=<secret-access-key> --bucket=<bucket>
dpl --provider=gcs --access-key-id=<access-key-id> --secret-access-key=<secret-access-key> --bucket=<bucket> --local-dir= BUILD
dpl --provider=gcs --access-key-id=<access-key-id> --secret-access-key=<secret-access-key> --bucket=<bucket> --acl=public-read
dpl --provider=gcs --access-key-id=<access-key-id> --secret-access-key=<secret-access-key> --bucket=<bucket> --detect-encoding --cache_control=max-age=99999
dpl --provider=gcs --access-key-id=<access-key-id> --secret-access-key=<secret-access-key> --bucket=<bucket> --local-dir=BUILD --upload-dir=BUILDS
- access_token: Optional. The access_token which can be found in the
.bitballoon
file after a deployment using the bitballoon CLI. Only required if no.bitballoon
file is present. - site_id: Optional. The site_id which can be found in the .bitballoon file after a deployment using the bitballoon CLI. Only required if no
.bitballoon
file is present. - local_dir: Optional. The sub-directory of the built assets for deployment. Default to current path.
dpl --access-token=<access-token> --site-id=3f932c1e-708b-4573-938a-a07d9728c22e
dpl --access-token=<access-token> --site-id=3f932c1e-708b-4573-938a-a07d9728c22e --local-dir=build
- user: Required. The user name at Puppet forge.
- password: Required. The Puppet forge password.
- url: Optional. The forge URL to deploy to. Defaults to https://forgeapi.puppetlabs.com/
dpl --provider=puppetforge --user=puppetlabs --password=s3cr3t
- username: Required. The packagecloud.io username.
- token: Required. The packagecloud.io api token.
- repository: Required. The repository to push to.
- local_dir: Optional. The sub-directory of the built assets for deployment. Default to current path.
- dist: Required for deb and rpm. The complete list of supported strings can be found on the packagecloud.io docs
dpl --provider=packagecloud --username=packageuser --token=t0k3n --repository=myrepo
dpl --provider=packagecloud --username=packageuser --token=t0k3n --repository=myrepo --dist=ubuntu/precise
dpl --provider=packagecloud --username=packageuser --token=t0k3n --repository=myrepo --local-dir="${TRAVIS_BUILD_DIR}/pkgs" --dist=ubuntu/precise
- target: Required. The git remote repository to deploy to.
- path: Optional. If using the skip_cleanup option to deploy from current file state, you can optionally specify the pathspec for the files to deploy. If not specified then all files are deployed.
dpl --provider=catalyze --target=ssh://git@git.catalyzeapps.com:2222/app1234.git
dpl --provider=catalyze --target=ssh://git@git.catalyzeapps.com:2222/app1234.git --skip_cleanup=true
dpl --provider=catalyze --target=ssh://git@git.catalyzeapps.com:2222/app1234.git --skip_cleanup=true --path=build
- Get the deployment target for Catalyze:
a. Make sure your catalyze environment is associated.
b. Get the git remote by running
git remote -v
from within the associated repo. - Setup a deployment key to Catalyze for Travis CI:
a. Install the travis-ci cli.
b. Get the public SSH key for your travis project and save it to a file by running
travis pubkey > travis.pub
c. Add the key as a deploy key using the catalyze cli within the associated repo. For example:catalyze deploy-keys add travisci ./travis.pub code-1
- Setup Catalyze as a known host for Travis CI:
a. List your known hosts by running
cat ~/.ssh/known_hosts
b. Find and copy the line from known_hosts that includes the git remote found in step #1. It'll look something like "[git.catalyzeapps.com]:2222 ecdsa-sha2-nistp256 BBBB12abZmKlLXNo..." c. Update yourbefore_deploy
step in.travis.yml
to update theknown_hosts
file:
before_deploy: echo "[git.catalyzeapps.com]:2222 ecdsa-sha2-nistp256 BBBB12abZmKlLXNo..." >> ~/.ssh/known_hosts
- user_id: Required. The user name at Chef Supermarket.
- client_key: Required. The client API key file name.
- cookbook_category: Required. The cookbook category in Supermarket (see: https://docs.getchef.com/knife_cookbook_site.html#id12 )
dpl --provider=chef-supermarket --user-id=chef --client-key=.travis/client.pem --cookbook-category=Others
- function_name: Required. The name of the Lambda being created / updated.
- role: Required. The ARN of the IAM role to assign to this Lambda function.
- handler_name: Required. The function that Lambda calls to begin execution. For NodeJS, it is exported function for the module.
- module_name: Optional. The name of the module that exports the handler. Defaults to
index
. - zip: Optional. Either a path to an existing packaged (zipped) Lambda, a directory to package, or a single file to package. Defaults to
Dir.pwd
. - description: Optional. The description of the Lambda being created / updated. Defaults to "Deploy build #{context.env['TRAVIS_BUILD_NUMBER']} to AWS Lambda via Travis CI"
- timeout: Optional. The function execution time at which Lambda should terminate the function. Defaults to 3 (seconds).
- memory_size: Optional. The amount of memory in MB to allocate to this Lambda. Defaults to 128.
- runtime: Optional. The Lambda runtime to use. Defaults to
node
.
Deploy contents of current working directory using default module:
dpl --provider="lambda" \
--access_key_id="${AWS_ACCESS_KEY}" \
--secret_access_key="${AWS_SECRET_KEY}" \
--function_name="test-lambda" \
--role="${AWS_LAMBDA_ROLE}" \
--handler_name="handler";
Deploy contents of a specific directory using specific module name:
dpl --provider="lambda" \
--access_key_id="${AWS_ACCESS_KEY}" \
--secret_access_key="${AWS_SECRET_KEY}" \
--function_name="test-lambda" \
--role="${AWS_LAMBDA_ROLE}" \
--zip="${TRAVIS_BUILD_DIR}/dist" \
--module_name="copy" \
--handler_name="handler";
- slug: Required.
~user-name/project-name/branch-name
- oauth_token: Required. Your OAUTH token for Launchpad
- oauth_token_secret: Required. Your OAUTH token secret for Launchpad
Deploy contents of current working directory using default module:
dpl --provider="launchpad" \
--slug="~user-name/project-name/branch-name" \
--oauth_token="${LAUNCHPAD_OAUTH_TOKEN}" \
--oauth_token_secret="${LAUNCHPAD_OAUTH_TOKEN_SECRET}";
Your Android(apk)/iOS(ipa) file will be uploaded to TestFairy, and your testers can start testing your app.
- api-key: TestFairy API Key (https://app.testfairy.com/settings/) run "travis encrypt --add deploy.api-key" on your repo.
- app-file: Path to the app file that will be generated after the build (APK/IPA).
- symbols-file: Path to the symbols file.
- keystore-file: Path to your keystore-file (must, only for android). http://docs.travis-ci.com/user/encrypting-files/
- storepass: storepass (must, only for android).
- alias: alias (must, only for android).
- testers-groups: You can set a tester group to be notified about this build (group1,group1).
- notify: If true, an email you a changelog will be sent to your users.
- auto-update: If true, all the previous installations of this app will be automatically all upgraded to this version.
- video-quality: Video quality settings, "high", "medium" or "low". Default is "high".
- screenshot-interval: You can choose "1""2""10" sec.
- max-duration: Maximum session recording length, eg "20m" or "1h". Default is "10m". Maximum "24h".
- advanced-options: For example (option1,option2)
- data-only-wifi: If true, video and recorded metrics will be sent only when connected to a wifi network.
- record-on-background: If true, data will be collected while the app on background.
- video: If true, Video recording settings "true", "false". Default is "true".
- icon-watermark: Add a small watermark to app icon. Default is "false".
- metrics: Comma-separated list of metrics to record. View list on http://docs.testfairy.com/Upload_API.html.
dpl --provider=testfairy --api-key=<api-key> --app-file="out/Sample.apk" --keystore-file="out/keystore" --storepass=<storepass> --alias=<alias>
- access-key-id: AWS Access Key.
- secret_access_key: AWS Secret Access Key.
- application: CodeDeploy Application Name.
- deployment_group: CodeDeploy Deployment Group Name.
- revision_type: CodeDeploy Revision Type (S3 or GitHub).
- commit_id: Commit ID in case of GitHub.
- repository: Repository Name in case of GitHub.
- region: AWS Availability Zone.
- wait-until-deployed: Wait until the app is deployed and return the deployment status (Optional, Default false).
- AWS_ACCESS_KEY_ID: AWS Access Key ID. Used if the
access-key-id
option is omitted. - AWS_SECRET_ACCESS_KEY: AWS Secret Key. Used if the
secret-access-key
option is omitted.
dpl --provider=codedeploy --access-key-id=<aws access key> --secret_access_key=<aws secret access key> --application=<application name> --deployment_group=<deployment group> --revision_type=<s3/github> --commit_id=<commit ID> --repository=<repo name> --region=<AWS availability zone> --wait-until-deployed=<true>
- email: ExoScale email or Organization ID.
- password: ExoScale password.
- deployment: ExoScale Deployment. Follows the format "APP_NAME/DEP_NAME".
dpl --provider=exoscale --email=<email> --password<password> --deployment=`APP_NAME/DEP_NAME`
- api_key: scalingo API Key. Not necessary if username and password are used.
- username: scalingo username. Not necessary if api_key is used.
- password: scalingo password. Not necessary if api_key is used.
- remote: Remote url or git remote name of your git repository. By default remote name is "scalingo".
- branch: Branch of your git repository. By default branch name is "master".
- app: Only necessary if your repository does not contain the appropriate remote. Specifying the app will add a remote to your local repository:
git remote add <remote> git@scalingo.com:<app>.git
You can connect to Scalingo using your username/password or your api key. It needs Scalingo CLI which will be downloaded here. Then, it will push your project to Scalingo and deploy it automatically.
Note: You only need to connect once to Scalingo CLI, credentials are stored locally.
dpl --provider=scalingo --api_key="aaAAbbBB0011223344"
dpl --provider=scalingo --username=<username> --password=<password>
dpl --provider=scalingo --api_key="aaAAbbBB0011223344" --remote="scalingo-staging"
dpl --provider=scalingo --api_key="aaAAbbBB0011223344" --remote="scalingo-staging" --branch="master"
dpl --provider=scalingo
An elementary provider that executes a single command.
Deployment will be marked a failure if the script exits with nonzero status.
- script: script to execute.
dpl --provider=script --script=./script/deploy.sh
Deploys to Google App Engine and Google App Engine Managed VMs via the Google Cloud SDK and
it's gcloud
tool using a Service Account.
In order to use this provider, please make sure you have the App Engine Admin API enabled in the Google Developers Console.
- project: Project ID used to identify the project on Google Cloud.
- keyfile: Path to the JSON file containing your Service Account credentials in JSON Web Token format. To be obtained via the Google Developers Console. Defaults to
"service-account.json"
. Note that this file should be handled with care as it contains authorization keys. - config: Path to your module configuration file. Defaults to
"app.yaml"
. This file is runtime dependent (Go, Java, PHP, Python) - version: The version of the app that will be created or replaced by this deployment. If you do not specify a version, one will be generated for you. See
gcloud preview app deploy
- no_promote: Flag to not promote the deployed version. See
gcloud preview app deploy
- verbosity: Let's you adjust the verbosity when invoking
"gcloud"
. Defaults to"warning"
. Seegcloud
. - no_stop_previous_version: Flag to prevent your deployment from stopping the previously promoted version. This is from the future, so might not work (yet). See
gcloud preview app deploy
- GOOGLECLOUDPROJECT or CLOUDSDK_CORE_PROJECT: Can be used instead of the
project
option. - GOOGLECLOUDKEYFILE: Can be used instead of the
keyfile
option.
dpl --provider=gae --project=example --no_promote=true
- token: Your Firebase CI access token (generate with
firebase login:ci
) - project: Deploy to a different Firebase project than specified in your
firebase.json
(e.g.myapp-staging
)
dpl --provider=firebase --token=<token> --project=<project>
- project Path to project folder relative to repo root. Defaults to repo root if not set.
- domain Domain to publish to. Can be omitted if domain is set in the
CNAME
file in the project folder.
- SURGE_LOGIN: Set it to the email address you use with Surge
- SURGE_TOKEN: Set it to your login token (get it by doing a
surge token
)
dpl --provider=surge --project=<project-path> --domain=<domain-name>