Issues
- 1
Create correct outputs from Falcon by changing the generation configuration
#103 opened by sebastian-weisshaar - 1
Add support for deep speed
#119 opened by alaeddine-13 - 0
- 0
Add automatic evaluation with gpt3
#97 opened by alaeddine-13 - 0
- 0
support codegen 1B in our training
#120 opened by alaeddine-13 - 0
Python QA instruction tuning dataset
#77 opened by JohannesMessner - 1
Update HF models
#93 opened by samsja - 0
Create evaluation harness with ChatGPT
#118 opened by sebastian-weisshaar - 0
Support mosaicml dolly_hhrlhf dataset
#114 opened by alaeddine-13 - 1
Align falcon 40b on code alpaca
#32 opened by samsja - 0
fixing transformers version
#87 opened by samsja - 1
add redpajama 7b to our pipeline
#48 opened by samsja - 2
Align Falcon 7b on Lima
#80 opened by sebastian-weisshaar - 0
Align falcon 7b on alpaca
#34 opened by samsja - 0
- 1
There are cases where the model is not stopping or repeats itself. We will try training for longer and see what happens
#73 opened by alaeddine-13 - 0
Add dolly 15k instruction dataset
#96 opened by alaeddine-13 - 0
- 0
We need to train alpaca-lora on the same number of lora layers to be able to compare it to falcon 7B and understand the effect of changing from llama to falcon
#71 opened by alaeddine-13 - 0
Align Falcon 7B on Lima
#79 opened by sebastian-weisshaar - 0
For Falcon, there are cases where the generation outputs an EOS token but does not stop
#72 opened by alaeddine-13 - 1
Add lima dataset to the training pipeline
#64 opened by alaeddine-13 - 0
Align Falcon 40b on alpaca-lora
#85 opened by sebastian-weisshaar - 0
Experiment with Lightning fabric, reproduce speed improvement from: https://lightning.ai/pages/community/finetuning-falcon-efficiently/
#86 opened by sebastian-weisshaar - 0
Compare code aligned model to current SOTA
#70 opened by azayz - 0
Fix bug in save_pretrained
#59 opened by samsja - 0
Add dockerfile to jerboa to run on runpod
#53 opened by samsja - 0
- 0
Pipeline training dataset refactoring
#51 opened by alaeddine-13 - 0
logs dataset in wandb
#52 opened by samsja - 1
Align llama7 on code alpaca
#33 opened by samsja - 1
Add qlora to our current codebase
#22 opened by samsja - 0
Allign llama 7b on alpaca with 4 bits
#31 opened by samsja - 8
Long term: create a good evaluation for QA code
#35 opened by samsja - 1
Publish alpaca lora 8 bits on our HF account
#36 opened by samsja - 0
- 0
Fix evaluation OOM
#12 opened by JohannesMessner - 0
Paper False promise llm
#21 opened by samsja - 1
WandB: Upload artifacts
#11 opened by JohannesMessner - 1
WandB: Remove unwante loss chart
#13 opened by JohannesMessner - 0
Create a tiny LLama model to run test
#1 opened by samsja