Daniel Fried*, Armen Aghajanyan*, Jessy Lin, Sida Wang, Eric Wallace, Freda Shi, Ruiqi Zhong, Wen-tau Yih, Luke Zettlemoyer, and Mike Lewis
This repository hosts example code showing how to use the model using HuggingFace's transformers
library. Code to replicate the evaluation results in our paper (in Fairseq, which we used to train the model) is coming soon!
See our project site for more information, or our paper, or examples.
You can obtain the models from HuggingFace's hub:
- 6.7B parameter model: facebook/incoder-6b
- 1.3B parameter model: facebook/incoder-1b
pytorch
, tokenizers
, and transformers
.
Our model requires HF's tokenizers >= 0.12.1, due to changes in the pretokenizer.
pip install pytorch
pip install 'tokenizers>=0.12'
pip install transformers
See example_usage.py for a demo script showing how to use the infilling capability of the model. Set BIG_MODEL = True in the script to use the 6.7B parameter model; the 1.3B will be used otherwise.
See our paper for research details on the method, training data, models, and experimental results.
See a demo of the 6.7B model on HF Spaces.
CC-BY-NC 4.0
Thanks to Lucile Saulnier, Leandro von Werra, Nicolas Patry, Suraj Patil, Omar Sanseviero, and others at HuggingFace for help with the model release, and to Naman Goyal and Stephen Roller for the code our demo was based on!