Issues
- 0
- 1
Whisper olive setup error
#1312 opened by liuyulvv - 0
- 4
- 0
capture-onnx-graph CLI Bug: list append()
#1348 opened by samuel100 - 1
Mistral int4 error
#1330 opened by eddan168 - 7
- 1
Mistral optimization(GPU) for a locally saved model, Failed to run Olive on gpu-cuda.
#1341 opened by tjinjin95 - 0
- 6
- 1
KeyError: 'unet_dataloader' occurs when optimizing unet in stable_diffusion_xl.py
#1327 opened by giocafe - 14
Whisper with DirectML EP not working: [ONNXRuntimeError] : 9 : NOT_IMPLEMENTED : Could not find an implementation for DecoderMaskedMultiHeadAttention(1) node with name 'Attention_0'
#1213 opened by WA225 - 4
- 0
error while inferencing the mistral LLM
#1309 opened by himanshushukla12 - 3
[FR]: Could not find a version that satisfies the requirement ort-nightly-directml==1.18.0 (from version: none)
#1280 opened by purejomo - 1
Missing implementation error for CoreML
#1299 opened by thewh1teagle - 2
whisper transcriptions is empty
#1291 opened by thewh1teagle - 1
"num_images" doesn't work for the example of directml stable_diffusion_xl.
#1164 opened by peterer0625 - 0
Very slow inference of optimized whisper gpu
#1300 opened by thewh1teagle - 0
Optimize whisper medium gpu failed
#1298 opened by thewh1teagle - 0
Whisper with DirectML: [ONNXRuntimeError] : 6 : RUNTIME_EXCEPTION : Non-zero status code returned while running WhisperBeamSearch node
#1221 opened by WA225 - 3
LLM Optimization with DirectML reply only displays "O"s
#1282 opened by yichunx1 - 1
Getting KeyError: 'input_model' when trying to optimize whisper-tiny.en model
#1283 opened by MayuraRam - 1
Get Error while optimizing SDXL of DirectML example
#1165 opened by klin2024 - 6
vae_encoder_gpu-dml_footprints.json file not found when converting stable diffusion xl base model
#1202 opened by AshD - 5
getting error while running llama2/bert on GPU
#1279 opened by himanshushukla12 - 1
- 2
- 2
Whisper optimization using ORT toolchain
#1264 opened by reeselevine - 2
[FR]: Gather per-pass output logs
#1269 opened by skywall - 1
- 6
Exporting to ONNX with kv_cache
#1240 opened by idruker-cerence - 0
nice pcode
#1248 opened by chemitax202 - 3
optimizing whisper without audio_decoder
#1236 opened by Aiurus - 6
[bug] Failed to perform inference on the optimized distil-large-v3 using the test_transcription.py
#1231 opened by MrRace - 3
- 4
[Whisper] convert with FP32 running failed
#1207 opened by Jay19751103 - 2
Using Whisper for Chinese ASR in iOS may occasionally output illegal UTF-8 strings.
#1197 opened by hasayakey - 1
- 0
Failed to optimize Whisper tiny by gpu (CUDA)
#1196 opened by XciciciX - 3
- 0
- 2
Llava-7b model Conversion to ONNX and Latency Optimization - OOM error (even after setting paging file size)
#1144 opened by Harini-Vemula-2382 - 1
[FR]: Finetuning QLoRA adapters and swapping
#1182 opened by franperezlopez - 0
onnx
#1156 opened by xalteropsx - 0
safetensor model
#1153 opened by neonarc4 - 0
huggingface_hub.errors.HFValidationError: Repo id must use alphanumeric chars or '-', '_', '.', '--' and '..' are forbidden, '-' and '.'
#1158 opened by neonarc4 - 2
Vitis quantization is broken with ORT 1.18
#1125 opened by guotuofeng - 2
Enabling openai/whisper-large-v3 using olive-ai-0.6.0 [onnxruntime-gpu: 1.17.1] on Intel CPU/GPU is not supporting
#1134 opened by vijayaVTT - 1