title | emoji | colorFrom | colorTo | sdk | sdk_version | app_file | pinned |
---|---|---|---|---|---|---|---|
RLHF |
🏢 |
red |
gray |
gradio |
3.1 |
app.py |
false |
An RLHF interface for data collection with Amazon Mechanical Turk and Gradio.
First, create a Python virtual environment and install the project's dependencies as follows:
python -m pip install -r requirements.txt
- Clone this repo and deploy it on your own Hugging Face space.
- Add the following secrets to your space:
HF_TOKEN
: One of your Hugging Face tokens.DATASET_REPO_URL
: The url to an empty dataset that you created the hub. It can be a private or public dataset.FORCE_PUSH
: "yes" When you run this space on mturk and when people visit your space on huggingface.co, the app will use your token to automatically store new HITs in your dataset. SettingFORCE_PUSH
to "yes" ensures that your repo will force push changes to the dataset during data collection. Otherwise, accidental manual changes to your dataset could result in your space getting merge conflicts as it automatically tries to push the dataset to the hub. For local development, add these three keys to a.env
file, and consider settingFORCE_PUSH
to "no".
To launch the Space locally, run:
python app.py
The app will then be available at a local address, such as http://127.0.0.1:7860
- On your local repo that you pulled, create a copy of
config.py.example
, just calledconfig.py
. Now, put keys from your AWS account inconfig.py
. These keys should be for an AWS account that has the AmazonMechanicalTurkFullAccess permission. You also need to create an mturk requestor account associated with your AWS account. - Run
python collect.py
locally.
Now, you should be watching hits come into your Hugging Face dataset automatically!
- Use caution while doing local development of your space and
simultaneously running it on mturk. Consider setting
FORCE_PUSH
to "no" in your local.env
file. - huggingface spaces have limited computational resources and memory. If you run too many HITs and/or assignments at once, then you could encounter issues. You could also encounter issues if you are trying to create a dataset that is very large. Check the log of your space for any errors that could be happening.