epfml/landmark-attention
Landmark Attention: Random-Access Infinite Context Length for Transformers
PythonApache-2.0
Issues
- 0
Fine-tuned Weights Doesn't Work
#16 opened by baotruyenthach - 0
- 3
Question about training stability
#14 opened by meme-virus - 0
- 36
- 1
- 2
- 0
Peft module
#10 opened by NicolasMejiaPetit - 0
How much VRAM do you need to run Inference?
#9 opened by FFFiend - 3
- 2
- 0
- 0
hello! i reached out on twitter before the release and received the link here when you guys dropped this
#5 opened by Alignment-Lab-AI - 1
Readme clarification
#3 opened by StrangeTcy - 0
model_name_or_path defaults
#4 opened by StrangeTcy