Make continuous deployment safe by comparing before and after webpage screenshots for each release. Depicted shows when any visual, perceptual differences are found. This is the ultimate, automated end-to-end test.
Depicted is:
- An API server for capturing webpage screenshots and automatically generating visual, perceptual difference images ("pdiffs").
- A workflow for teams to coordinate new releases using pdiffs.
- A client library for integrating with existing continuous integration processes.
- Built for portability; API server runs on App Engine, behind the firewall, etc.
- A wrapper of PhantomJS for screenshots.
- Open source, Apache 2.0 licensed.
- Not a framework, not a religion.
Depicted is not finished! Please let us know if you have feedback or find bugs.
See this video for a presentation about how perceptual diffs have made continuous deployment safe.
Here are the steps to making Depicted useful to you:
- Establish a baseline release with an initial set of screenshots of your site.
- Create a new release with a new set of screenshots of your new version.
- Manually approve or reject each difference the tool finds.
- Manually mark the new release as good or bad.
- Repeat. Your approved release will become the baseline for the next one.
Depicted organizes your releases by a build ID. You can create a build through the API server's UI. A build is usually synonymous with a binary that's pushed to production. But it could also be a unique view into one product, like one build for desktop web and another for mobile web.
Within a build are releases with names. I recommend naming a release as the day it was branched in your source repository, and maybe an attempt number, like "06-16-r01" for June 16th, release 1. If you use codenames for your releases, like "bumblebee", that works too.
Each release may be attempted many times. The full history of each release attempt is saved in the UI. Releases can be manually marked as good or bad in the UI. When results for a new release are uploaded, they will automatically be compared to the last known-good version within that build.
A release consists of many separate test runs. A test run is a single screenshot of a single page. A test run has a name that is used to pair it up with a baseline test run from the known-good, previous release. Usually the test run is named as the path of the URL being tested (like /foo?bar=meep). This lets the baseline release and new release serve on different hostnames.
The life-cycle of a release:
- Created: A new release is created with a specific name. The system gives it a release number.
- Receiving: The release is waiting for all test runs to be requested or reported.
- Processing: All test runs have been reported, but additional processing (like screenshotting or pdiffing) is required.
- Reviewing: All test runs have been processed. Now the build admin should review any pdiffs that were found and approve the release.
Final release states:
- Bad: The build admin has marked the release and all its test runs as bad. It will never be used as a baseline.
- Good: The build admin has marked the release and all of its test runs as passing. The next release created for this build will use this just-approved release as the new baseline.
You can try out the API on the test instance of Depicted located at https://dpxdt-test.appspot.com. This instance's database will be dropped from time to time, so please don't rely on it.
The API is really simple. All requests are POSTs with parameters that are URL encoded. All responses are JSON. All requests should be over HTTPS. The API server uses HTTP Basic Authentication to verify your client has access to your builds. You can provision API keys for a build on its homepage.
Here's an example request to the API server using curl. Pretty simple.
curl -v \
-u api_key:api_password \
-F build_id=1 \
-F 'run_name=/static/dummy/dummy_page1.html' \
-F 'release_number=1' \
-F 'log=906d3259c103f6fcba4e8164a4dc3ae0d1a685d9' \
-F 'release_name=2013-06-16 17:35:03.327710' \
'http://localhost:5000/api/report_run'
An example client tool that exercises the whole workflow is available in the client repository here. It's called "Site Diff". It will crawl a webpage, follow all links with the same prefix path, then create a new release that screenshots all the URLs. Running the tool multiple times lets you diff your entire site with little effort. Site Diff is very helpful, for example, when you have a blog with a lot of content and want to make a change to your base template and be sure you haven't broken any pages.
Here's an example invocation of site_diff:
./site_diff.py \
--phantomjs_binary=path/to/phantomjs-1.8.1-macosx/bin/phantomjs \
--phantomjs_script=path/to/client/capture.js \
--upload_build_id=1234 \
--release_server_prefix=https://my-dpxdt-apiserver.example.com/api \
http://www.example.com/my/website/here
All of these requests are POSTs with URL-encoded or multipart/form-data bodies and require HTTP Basic Authentication using your API key as the username and secret as the password. All responses are JSON. The 'success' key will be present in all responses and true if the request was successful. If 'success' isn't present, a human-readable error message may be present in the response under the key 'error'.
Creates a new release candidate for a build.
- build_id
- ID of the build.
- release_name
- Name of the new release.
- url
- URL of the homepage of the new release. Only present for humans who need to understand what a release is for.
- build_id
- ID of the build.
- release_name
- Name of the release that was just created.
- release_number
- Number assigned to the new release by the system.
- url
- URL of the release's homepage.
Finds the last good run of the given name for a release. Returns an error if no run previous good release exists.
- build_id
- ID of the build.
- run_name
- Name of the run to find the last known-good version of.
- build_id
- ID of the build.
- release_name
- Name of the last known-good release for the run.
- release_number
- Number of the last known-good release for the run.
- run_name
- Name of the run that was found. May be null if a run could not be found.
- url
- URL of the last known-good release for the run. May be null if a run could not be found.
- image
- Artifact ID (SHA1 hash) of the screenshot image associated with the run. May be null if a run could not be found.
- log
- Artifact ID (SHA1 hash) of the log file from the screenshot process associated with the run. May be null if a run could not be found.
- config
- Artifact ID (SHA1 hash) of the config file used for the screenshot process associated with the run. May be null if a run could not be found.
Requests a new run for a release candidate. Causes the API system to take screenshots and do pdiffs.
- build_id
- ID of the build.
- release_name
- Name of the release.
- release_number
- Number of the release.
- url
- URL to request as a run.
- config
- JSON data that is the config for the new run.
The config passed to the request_run
function may have any or all of these fields. All fields are optional and have reasonably sane defaults.
{
"viewportSize": {
"width": 1024,
"height": 768
},
"injectCss": ".my-css-rules-here { display: none; }",
"injectJs": "document.getElementById('foobar').innerText = 'foo';"
}
- build_id
- ID of the build.
- release_name
- Name of the release.
- release_number
- Number of the release.
- run_name
- Name of the run that was created.
- url
- URL that was requested for the run.
- config
- Artifact ID (SHA1 hash) of the config file that will be used for the screenshot process associated with the run.
Uploads an artifact referenced by a run.
- build_id
- ID of the build.
- (a single file in the multipart/form-data)
- Data of the file being uploaded. Should have a filename in the mime headers so the system can infer the content type of the uploaded asset.
- build_id
- ID of the build.
- sha1sum
- Artifact ID (SHA1 hash) of the file that was uploaded.
- content_type
- Content type of the artifact that was uploaded.
Reports data for a run for a release candidate. May be called multiple times as progress is made for a run. No longer callable once the screenshot image for the run has been assigned.
- build_id
- ID of the build.
- release_name
- Name of the release.
- release_number
- Number of the release.
- run_name
- Name of the run.
- url
- URL associated with the run.
- image
- Artifact ID (SHA1 hash) of the screenshot image associated with the run.
- log
- Artifact ID (SHA1 hash) of the log file from the screenshot process associated with the run.
- config
- Artifact ID (SHA1 hash) of the config file used for the screenshot process associated with the run.
- ref_url
- URL associated with the run's baseline release.
- ref_image
- Artifact ID (SHA1 hash) of the screenshot image associated with the run's baseline release.
- ref_log
- Artifact ID (SHA1 hash) of the log file from the screenshot process associated with the run's baseline release.
- ref_config
- Artifact ID (SHA1 hash) of the config file used for the screenshot process associated with the run's baseline release.
- diff_image
- Artifact ID (SHA1 hash) of the perceptual diff image associated with the run.
- diff_log
- Artifact ID (SHA1 hash) of the log file from the perceptual diff process associated with the run.
- diff_success
- Present and non-empty string when the diff process ran successfully. May be missing when diff ran and reported a log but may need to retry for this run.
Nothing but success on success.
Marks a release candidate as having all runs reported.
- build_id
- ID of the build.
- release_name
- Name of the release.
- release_number
- Number of the release.
- results_url
- URL where a release candidates run status can be viewed in a web browser by a build admin.
Depicted is written in portable Python. It uses Flask and SQLAlchemy to make it easy to run in your environment. It works with SQLite out of the box. The API server runs on App Engine. The workers run ImageMagick and PhantomJS as subprocesses. I like to run the worker on a cloud VM, but you could run it on your laptop behind a firewall if that's important to you.
Update the common.sh file to match your environment.
To use the server locally for the first time, or when the schema changes during development:
# ./run_shell.sh
>>> from dpxdt.server import db
>>> db.drop_all()
>>> db.create_all()
Create a secrets.py file in the root of the project with these keys:
# Please generate a reasonable key. Used by Flask for CSRF, Login cookie, etc.
SECRET_KEY = 'my-key-here'
To run the API server locally, without any worker threads:
./run_server.sh
To run the background workers independently against the local API server:
./run_worker.sh
To run the API server locally with all background workers:
./run_combined.sh
To run in the App Engine development environment:
./run_appengine.sh
This is rough. Primarily explains how to deploy to App Engine / CloudSQL / Google Compute Engine.
Provision a CloudSQL DB for your project and initialize it:
./google_sql.sh dpxdt-cloud:test
sql> create database test;
Go to the Google API console and provision a new project and "API Access". This will give you the OAuth client ID and secret you need to make auth work properly. Update config.py with your values.
Go to the deployment/test-appengine directory. Update app.yaml with your parameters. Create the secrets.py file as explained for development.
Deploy the app:
./deploy.sh
Navigate to /admin on your app and run in the interactive console:
from dpxdt import server
server.db.create_all()
Navigate to / on your app and see the homepage. Create a new build. Provision an API key. Then set your user and API key as superusers using the SQL tool:
select * from user;
update user set superuser = 1 where user.id = 'foo';
select * from api_key;
update api_key set superuser = 1 where id = 'foo';
Now create the background workers package to deploy from the deployment/test-worker directory:
./package.sh
Follow the commands it prints out to deploy the worker to a VM.
Nothing yet!