Issues
- 2
- 4
- 4
llava support
#88 opened by sonic182 - 6
- 0
[Question] BasicTransformerBlock
#96 opened by JH-ninjatech - 5
- 2
- 8
NaN outputs when masking llama model inputs
#79 opened by dacorvo - 1
Vicuna13B model support
#66 opened by petrovicu - 6
Infering logits from `model.forward` for the entire batch instead of the last forward's output.
#73 opened by michaelfeil - 3
- 4
Can't save/serialize any models except GPT2
#58 opened by awskila - 0
Any plan to support Qwen-2 Model
#89 opened by mynewstart - 1
Add support for `gemma` models
#82 opened by benglewis - 4
Compilation error on llama 7 B with batch size 8
#59 opened by dacorvo - 7
- 4
Improve Neuron model loading time
#80 opened by dacorvo - 5
Generate Llama 2 from Embeddings
#72 opened by liechtym - 8
Mixtral config issue -- not handling null well
#71 opened by jimburtoft - 0
Add support for Baichuan-13B model
#83 opened by cszhz - 2
- 6
- 5
Issue while compiling Mistral 7B 0.2 Instruct
#77 opened by josete89 - 4
Support for Mistral-7B model
#50 opened by henghui-zhu-amazon - 5
Possible error in top-p filtering
#46 opened by dacorvo - 3
save_split seems to be broken after transformers made safetensor serialization default
#55 opened by jitto - 4
`stopping_criteria_list(input_ids, probs)` does not check for the correct sequence.
#75 opened by michaelfeil - 1
- 1
- 1
Support for MPT model
#74 opened by klutzDrawers - 3
- 3
Any solution to save the converted model?
#29 opened by aliseyfi - 11
Inf2 Modified Llama 2 Loading Issue
#67 opened by liechtym - 2
How to use generate() with inputs_embeds
#70 opened by liechtym - 8
Compilation errors for llama 2 models
#45 opened by dacorvo - 2
Support for encoder-decoder models
#51 opened by kwontaek-amazon - 2
Mixtral Model support
#65 opened by enochlev - 4
llama-2/codellama benchmark for inf2.xlarge
#64 opened by zliendo - 6
Llama2 inference overhead time way too long
#63 opened by enochlev - 4
- 4
AssertionError when running fine-tuned llama 2
#40 opened by eladspi - 2
- 1
from_pretrained is broken after transformers made safetensor serialization default
#60 opened by dennj - 4
- 4
- 13
- 2
How to set: FI_EFA_FORK_SAFE=1 ?
#37 opened by yogendra-yatnalkar - 3
ImportError: cannot import name 'neuron_xla_compile' from 'libneuronxla'
#33 opened by yogendra-yatnalkar - 2
Corrupted output with llama prototype model
#30 opened by dacorvo - 2
neuronx-cc --target
#31 opened by sheenobu