We recommend using
- For 1B to 3B models, it's advisable to have at least NVIDIA T4, 10 Series, or 20 Series GPUs, or Apple Silicon like the M1.
- For 7B to 13B models, we recommend using NVIDIA V100, A100, 30 Series, or 40 Series GPUs.
We have published benchmarks for these models on https://leaderboard.tabbyml.com for Tabby's users to consider when making trade-offs between quality, licensing, and model size.
To ensure optimal response quality, and given that latency requirements are not stringent in this scenario, we recommend using a model with at least 1B parameters.
Model ID | License |
---|---|
Mistral-7B | Apache 2.0 |
CodeGemma-7B-Instruct | Gemma License |
Qwen2-1.5B-Instruct | Apache 2.0 |
CodeQwen-7B-Chat | Tongyi Qianwen License |
Codestral-22B | Mistral AI Non-Production License |
Model ID | License |
---|---|
Nomic-Embed-Text | Apache 2.0 |
Jina-Embeddings-V2-Code | Apache 2.0 |