Issues
- 3
- 0
nice pcode
#1248 opened - 8
Exporting to ONNX with kv_cache
#1240 opened - 3
optimizing whisper without audio_decoder
#1236 opened - 6
[bug] Failed to perform inference on the optimized distil-large-v3 using the test_transcription.py
#1231 opened - 3
- 0
- 14
- 4
[Whisper] convert with FP32 running failed
#1207 opened - 6
vae_encoder_gpu-dml_footprints.json file not found when converting stable diffusion xl base model
#1202 opened - 2
- 0
Failed to optimize Whisper tiny by gpu (CUDA)
#1196 opened - 0
- 3
- 1
- 1
[FR]: Finetuning QLoRA adapters and swapping
#1182 opened - 2
- 1
- 1
- 0
- 0
onnx
#1156 opened - 0
safetensor model
#1153 opened - 2
Llava-7b model Conversion to ONNX and Latency Optimization - OOM error (even after setting paging file size)
#1144 opened - 2
Enabling openai/whisper-large-v3 using olive-ai-0.6.0 [onnxruntime-gpu: 1.17.1] on Intel CPU/GPU is not supporting
#1134 opened - 2
Vitis quantization is broken with ORT 1.18
#1125 opened - 1
- 5
- 11
- 1
Exception while running SD XL: Not enough memory resources are available to complete this operation
#1079 opened - 17
- 1
- 1
[FR]: FlashAttention support for Whisper
#1065 opened - 2
Missing dependency: psutil
#1043 opened - 3
- 1
- 7
- 8
Whisper-medium conversion failed
#1023 opened - 5
- 4
models_rank.json issue
#1007 opened - 3
- 0
Conversion of some models are buggy
#939 opened - 4
This is an invalid model
#935 opened - 3
SDXL crashing when trying to run
#930 opened - 10
- 8
I don't have models/optimized/llama_v2 folder after I've run python llama_v2.py --optimize
#905 opened - 6
- 2
Error with search strategic.py 'Conversion Merged" has no output models for Llama2 optimization
#897 opened - 9
- 11
- 2
Missing file https://github.com/microsoft/Olive/blob/main/docs/source/overview/configuring_pass
#866 opened