abertsch72/unlimiformer
Public repo for the NeurIPS 2023 paper "Unlimiformer: Long-Range Transformers with Unlimited Length Input"
PythonMIT
Issues
- 4
BookSum_Full BART Baseline script/code
#66 opened by saxenarohit - 4
Unable to load dataset
#64 opened by Ozawa333 - 1
DatasetGenerationError
#65 opened by pppyb - 1
LLama2_example output random words
#60 opened by KerolosAtef - 6
Can't run the provided llama2 example
#59 opened by KerolosAtef - 0
Error in running Llama 2 generation example
#63 opened by OswaldHe - 2
- 0
How can we use unlimiformer for sequence classification (textual entailment)?
#62 opened by robinsingh-ai - 3
Running Unlimiformer with the `forward` method
#38 opened by testzer0 - 1
Script utilizing LLM
#51 opened by jcgeo9 - 1
GPU VRAM Usage during training
#58 opened by KevinD777 - 7
reproducing your results
#57 opened by patrickocal - 6
- 1
- 3
support other llms?
#34 opened by chaunceyliu30 - 1
Question: During training, the calculation of topk value’s att_weight is different from the classic transformer’s multi-head attention.
#45 opened by jjkk123456 - 3
About adding a prefix and input length
#47 opened by apapoudakis - 3
Why is the inference so slow?
#53 opened by cckao - 2
- 17
TypeError: torch_replacement_knn_gpu() got an unexpected keyword argument 'device'
#25 opened by jordancole21 - 1
- 5
Steps to run the code
#33 opened by sahulsumra - 3
- 8
Use of other Encode/Decoder Models
#55 opened by rdmerillat - 4
Why "import sled" was commented out in run.py?
#50 opened by shi-kejian - 0
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, ....
#49 opened by shi-kejian - 4
multi-gpu unlimiformer training: Expected all tensors to be on the same device
#52 opened by shi-kejian - 16
- 2
- 1
- 8
- 2
Why using different calculation methods for the key and value of the cross-attention of the decoder layer in the training and validation stages?
#44 opened by jjkk123456 - 2
API server for unlimiformer
#39 opened by neubig - 2
Set max_size to 128 but use 512 tokens
#43 opened by adivoj - 8
Errors on running llama with `test_datastore`
#41 opened by wywyWang - 2
error while training
#42 opened by kekekawaii2839 - 1
- 2
- 6
Sanity check: VRAM usage on llama-2-7b-chat-hf higher than without Unlimiformer on low tokens?
#26 opened by SharkWipf - 2
About the method `attention_forward_hook`
#30 opened by seunghyukoh - 1
- 1
- 2
Error while evaluating
#20 opened by MonliH - 7
Reproduce the +test Unlimiformer setup
#17 opened by Leonard907 - 2
Support multilingual model like mt0, mBart ?
#18 opened by trannhatquy - 5
Encoder Only Unlimiformer
#21 opened by YHL04 - 18
- 10
Working with 8bit and 4bit quantized models
#19 opened by jordancole21 - 1
Can unlimiformer be trained on mutiple gpus?
#16 opened by Muxv - 2
Typing checks fail
#13 opened by StrangeTcy