Pinned issues
Issues
- 4
Convert model.bin (fp32) to model.bin (int8)
#1761 opened - 0
- 4
- 3
Support for ARM64 on Windows
#1756 opened - 2
Docker images not published
#1754 opened - 2
CI failing in the several recent PRs
#1753 opened - 1
- 2
Llama 3.1 support please?
#1745 opened - 1
- 10
- 1
- 1
Introduce better format for whisper models
#1741 opened - 5
- 0
Falcon-11B support
#1737 opened - 1
[Feature Request] Expose Profiler to Python
#1736 opened - 2
Add gemma2 support
#1735 opened - 1
- 2
- 1
[feature request] Mixed quantizations.
#1730 opened - 3
Download ready to use model
#1729 opened - 4
Gemma model - help needed
#1728 opened - 4
Adding a layer to an existing model?
#1726 opened - 0
- 1
- 2
Qwen2 Support?
#1721 opened - 1
Does CT2/OpenNMT engine support Qualcomm SoC?
#1720 opened - 2
T5 inference result is all <pad>
#1719 opened - 18
- 0
- 7
- 1
Ctranslate2 Pypi exceeds limit 20GB
#1712 opened - 9
Converter not working for NLLB models
#1711 opened - 3
CUDA DeviceAllocate segfault
#1709 opened - 2
- 2
Support for Phi3-Small, Medium, and Vision
#1707 opened - 3
- 6
Doesn't build without docker. libiomp5 not found
#1703 opened - 6
Option --self_attn_type scaled-dot-flash is not supported (supported values are: scaled-dot)
#1702 opened - 1
- 0
- 6
CTranslate2 cmake error when trying to build the code from source with cuda support enabled on Windows.
#1697 opened - 2
- 0
opus-mt-en-zh does not respect the end token
#1694 opened - 5
Can't hide GPUs to get_cuda_device_count()
#1693 opened - 3
How to compile from source on windows 11?
#1692 opened - 4
- 2
target_prefix latency
#1689 opened - 4
[SOLVED] Running Llama3 with Ctranslate2
#1688 opened - 2
Dynamic LoRA switching
#1686 opened - 3