Issues
- 2
Guided choice not respected
#612 opened by Andrea-de-Varda - 1
RetNet adaptation
#266 opened by yunfeng-scale - 1
Further reduction of pod cold start time
#269 opened by yunfeng-scale - 1
PEFT adapters with continuous batching
#268 opened by yunfeng-scale - 1
Speculative decoding
#267 opened by yunfeng-scale - 1
Investigate CUDA graphs
#265 opened by yunfeng-scale - 1
GQA for Llama 2 7B and 13B models
#264 opened by yunfeng-scale - 1
Add support for Mistral in the FineTune API
#308 opened by ian-scale - 2
Return log probabilities of the prompt
#616 opened by ChiaSap - 0
⚠️ Notice: sunsetting the public demo of LLM Engine ⚠️
#619 opened by yixu34 - 1
Guided choice with mixtral model is not honored
#528 opened by KE7 - 1
Test out spot instances
#270 opened by yunfeng-scale - 1
how can we use glue dataset to llama2-7B?
#499 opened by ZHANGJINKUI - 1
Add Completions support for Llama 3
#497 opened by yixu34 - 1
Fine-Tuning LLM using Local GPU and Infra
#477 opened by Sree-abcprocure - 0
Integrate TensorRT-LLM
#363 opened by yunfeng-scale - 4
- 2
Add support for Mistral-7B in the Completions API
#301 opened by yixu34 - 1
self host on runnpod
#285 opened by Stealthwriter - 4
Error: Internal Server Error: <class 'AttributeError'>: 'CreateFineTuneResponse' object has no attribute 'artifact_id'
#272 opened by ashutoshrana171 - 3
Control frequency - completion
#277 opened by Stealthwriter - 4
⚠️ LLM Engine fine-tuning maintenance ⚠️
#233 opened by yixu34 - 2
Investigate Multi-Query Attention
#171 opened by yixu34 - 1
[Feature Request] support InternLM
#206 opened by JimmyMa99 - 0
Fine-tuning API should return immediately with a clear error message if the input is invalid
#213 opened by yixu34 - 0
Surface pandas ParserError to user
#184 opened by squeakymouse - 5
- 1
- 4
- 1
[Lora] Allow more Lora hyperparams
#163 opened by sam-scale - 1
[Datasets] s3 presigned url fails?
#168 opened by sam-scale - 1
Model ids =/= Fine Tune Id?
#167 opened by sam-scale - 1
Add github sidebar
#155 opened by yixu34 - 1
- 7
Llama-2-70B support
#166 opened by yixu34 - 9
- 1
[Tracking] Allow wandb tracking
#165 opened by sam-scale - 2
Import completion error
#182 opened by 4n9le-bot - 2
- 7
- 1
[Feature Request] Add `on_inference_ready` callback to llm-engine deployments via `Model.create`
#193 opened by jenkspt - 1
- 0
[dev] Deploy docs through CI
#189 opened by phil-scale - 3
- 0
Allow users to set API key without using env variables
#156 opened by yixu34 - 1
Example ScienceQA fine-tuning notebook has no documentation for how to install dependencies
#172 opened by rkaplan - 0
[Tokenizer] Allow for custom tokens/tokenizer
#164 opened by sam-scale - 0
- 7
model name parameter after finetuning
#143 opened by eshrager - 3
GKE Helm deployment
#157 opened by brosand