This is the first RESTful API for the Federal Election Commission. We're aiming to make campaign finance more accessible for journalists, academics, developers, and other transparency seekers. This is also fueling the campaign finance data in the upcoming FEC website.
Our code is licensed under CC0, but a few restrictions limit the way you can use FEC data. For example, you can’t use contributor lists for commercial purposes or to solicit donations. Learn more on FEC.gov.
Note: This project is still in alpha and not yet deployed. We're still investigating the best ways to present this data to the public.
Hello! If you're interested in learning more about this project, check out some related repos and don't be afraid to ask us questions (general questions are usually posted in the 18F/FEC repo).
If you'd like to contribute to our project, please check out our openfec repo. We try to tag things that are easy to pick up without being entrenched in our project with a "help wanted" tag. Things in our backlog are usually also up for grabs, so let us know if you'd like to pick something up from there.
If you are interested in contributing, please check out our contributing guidelines, which we use to guide our development processes internally. You don't have to adhere to them to participate, but following them may help keep things from getting messy.
If you would like to be an FEC API beta tester, get an API key, check out the experimental API, and give us your feedback by filing issues.
- FEC - A discussion forum for the project.
- openFEC - Where the API work happens. We also use this as the central repo to create issues related to each sprint and our backlog. If you're interested in contributing, please look for "help wanted" tags or ask!
- openFEC-web-app - Where the campaign finance web app work happens. Note that issues and discussion tend to happen in the other repos.
- fec-alpha - A place to explore and develop a new site for the Federal Election Commission.
The easiest way to get started with working on openFEC is to run the bootstrap script.
Prior to running, ensure you have the following requirements installed:
- virtualenv
- virtualenvwrapper
- python3.4
- pip
- nodejs
- npm
- PostgreSQL
- tmuxinator
Then, simply run:
$ curl https://raw.githubusercontent.com/18F/openFEC/master/scripts/bootstrap/fec_bootstrap.sh | bash
This will clone both openFEC repos, set up virtual environments, and set some environment variables (that you supply) in ~/.fec_vars. It might be a good idea to source that file in your ~/.bashrc or ~/.zshrc.
NOTE: This will also sync this repo. For bootstrapping, we recommend running the script prior to cloning the repo and letting the script handle that.
There is also a Vagrantfile and provisioning shell script available. This will create an Ubuntu 14.04 virtual machine, provisioned with all the requirements to run the bootstrap script.
From scripts/bootstrap, simply:
$ vagrant up
$ vagrant ssh
$ cp /vagrant/fec_bootstrap.sh fec_bootstrap.sh && ./fec_bootstrap.sh
Assuming you ran the bootstrap script, you can launch the API and the Web App with a single command:
$ tmuxinator fec-local
The site can be found at http://localhost:3000 (or http://localhost:3001 if using Vagrant). Remember the username and password you created when running the script.
To deploy to Cloud Foundry, run invoke deploy
. The deploy
task will attempt to detect the appropriate
Cloud Foundry space based the current branch; to override, pass the optional --space
flag:
$ invoke deploy --space dev
The deploy
task will use the FEC_CF_USERNAME
and FEC_CF_PASSWORD
environment variables to log in.
If these variables are not provided, you will be prompted for your Cloud Foundry credentials.
Credentials for Cloud Foundry applications are managed using user-provided services labeled as "fec-creds-prod", "fec-creds-stage", and "fec-creds-dev". Services are used to share credentials across blue and green versions of blue-green deploys, and between the API and the webapp. To set up a service:
$ cf target -s dev
$ cf cups fec-creds-dev -p '{"SQLA_CONN": "..."}'
To stand up a user-provided credential service that supports both the API and the webapp, ensure that the following keys are set:
- SQLA_CONN
- FEC_WEB_USERNAME
- FEC_WEB_PASSWORD
- FEC_WEB_API_KEY
- FEC_WEB_API_KEY_PUBLIC
- NEW_RELIC_LICENSE_KEY
Deploys of a single app can be performed manually by targeting the env/space, and specifying the corresponding manifest, as well as the app you want, like so:
$ cf target [dev|stage|prod] && cf push -f manifest_<[dev|stage|prod]>.yml [api|web]
The OpenFEC API is a Flask application deployed using the gunicorn WSGI server behind
an nginx reverse proxy. Static files are compressed and served directly through nginx;
dynamic content is routed to the Flask application via proxy_pass
. The entire application
is served through the API Umbrella, which handles API keys,
caching, and rate limiting.
All API responses are set to expire after one hour (Cache-Control: public, max-age=3600
).
In production, the API Umbrella will check this response header
and cache responses for the specified interval, such that repeated requests to a given
endpoint will only reach the Flask application once. This means that responses may be
stale for up to an hour following the nightly refresh of the materialized views.
The staging and production environments use the API Umbrella for
rate limiting, authentication, caching, and HTTPS termination and redirection. Both
environments use the FEC_API_WHITELIST_IPS
flag to reject requests that are not routed
through the API Umbrella.
We use git-flow for naming and versioning conventions. Both the API and web app are continuously deployed through Travis CI accordingly.
-
Developer creates a feature branch
$ git flow feature start my-feature
-
Reviewer merges feature branch into develop and pushes to origin
-
[auto] Develop is deployed to dev
-
Developer creates a hotfix branch
$ git flow hotfix start my-hotfix
-
Reviewer merges hotfix branch into develop and master and pushes to origin
-
[auto] Develop is deployed to dev
-
[auto] Master is deployed to prod
-
Developer creates a release branch and pushes to origin
$ git flow release start my-release $ git flow release publish my-release
-
[auto] Release is deployed to stage
-
Review of staging
-
Developer merges release branch into master and pushes to origin
$ git flow release finish my-release
-
[auto] Master is deployed to prod
Note: The following can be automated using Boto or the AWS CLI if we continue on with this model and need to update snapshots frequently.
The production and staging environments use RDS instances that receive streaming updates from the FEC database. The development environment uses a separate RDS instance created from a snapshot of the production instance. To update the development instance (e.g. when schemas change or new records are added):
-
Create a new snapshot of the production data
RDS :: Instances :: fec-goldengate-target :: Take DB Snapshot
-
Restore the snapshot to a new development RDS
RDS :: Snapshots :: <snapshot-name> :: Restore Snapshot
- DB Instance Class: db.m3.medium
- Multi-AZ Deployment: No
- Storage Type: General Purpose
- DB Instance Identifier: fec-goldengate-dev-YYYY-mm-dd
- VPC: Not in VPC
-
Add the new instance to the FEC security group
RDS :: Instances :: <instance-name> :: Modify
- Security Group: fec-open
- Apply Immediately
-
Configure DNS to point to new instance
Route 53 :: Hosted Zones :: open.fec.gov :: goldengate-dev.open.fec.gov
- Value:
- Example: fec-goldengate-dev-YYYY-mm-dd...rds.amazonaws.com
- Value:
-
Wait up to
TTL
seconds for DNS records to propagate -
Verify that new instance is reachable at goldengate-dev.open.fec.gov
-
Delete previous development instance
Important: Verify that all newly created instances are tagged with the same client as the production instance.
$ createdb cfdm_test
$ pg_restore --dbname cfdm_test data/subset.dump
$ ./manage.py update_all
$ nosetests
This repo includes a small subset of the staging database (built 2015/08/12) at data/subset.dump
. To use the test subset for local development:
$ pg_restore --dbname <dest> data/subset.dump
To build a new test subset, use the build_test
invoke task:
$ invoke build_test <source> <dest>
where both source
and dest
are valid PostgreSQL connection strings.
To update the version-controlled test subset after rebuilding, run:
$ invoke dump <source> data/subset.dump
where source
is the database containing the newly created test subset.
This repo includes optional post-merge and post-checkout hooks to ensure that
dependencies are up to date. If enabled, these hooks will update Python
dependencies on checking out or merging changes to requirements.txt
. To
enable the hooks, run
$ invoke add_hooks
To disable, run
$ invoke remove_hooks