Issues
- 1
Some observations and questions on Google FRAMES Benchmark readurls&memory-gpt-4o-mini method evaluation
#106 opened by RGSmirnov - 2
- 4
- 0
- 1
How to reproduce RTC Eval 100% locally?
#96 opened by botelhorui - 5
Implement cot decoding with llama.cpp
#65 opened by codelion - 0
parse conversation reports error
#84 opened by femto - 14
I get the following error: list index out of range
#67 opened by ErykCh - 1
Feature: easy to add new approaches
#72 opened by ErykCh - 6
Setting the default approach doesn't work
#69 opened by ErykCh - 6
Ambiguous configuration for mcts
#68 opened by ErykCh - 3
Resulting docker image size (6.36GB) is quite large - is there any opportunity to reduce this?
#71 opened by sammcj - 2
Thanks for adding the entropy based sampling, by any chance do you have a comprasion with other alternative methods?
#70 opened by shamanez - 6
(MOA) Fails with "List Index Out of Range" Error on OpenAI-Compatible Ollama API Endpoint
#60 opened by chrisoutwright - 1
Add a lighting template for running optillm
#56 opened by codelion - 1
Is there any possibility we align some interest?
#57 opened by femto - 1
Using llama-server issue with 'no_key' API key
#61 opened by s-hironobu - 1
Scripts to reproduce benchmark results
#63 opened by zhxieml - 1
- 1
Implement routing
#37 opened by codelion - 13
I can see cot_decode method has implemented, but we can't use it with the proxy.
#59 opened by shamanez - 6
When I tried the optillm with my own openai API compatible hosted model I get this error
#58 opened by shamanez - 2
Integration with Gemini 1.5 models
#54 opened by tranhoangnguyen03 - 2
token counting
#52 opened by darkacorn - 1
[Question]: Which paper is mcts.py based on?
#51 opened by RomanKoshkin - 20
- 2
- 1
Add support for logging with --log=debug
#44 opened by codelion - 1
Add support for sympy in solver approach
#41 opened by codelion - 1
- 0
Add support to pass slug as extra_body argument instead of prefix of model name
#39 opened by codelion - 4
- 7
- 2
Clarification: proxy or library for cot_decoding??
#35 opened by lee-b - 0
Change api-key to optillm-api-key
#24 opened by codelion - 8
use with llama.cpp
#8 opened by scalar27 - 5
Flask import fails
#23 opened by vanetreg - 0
- 1
Support AzureOpenAI client
#13 opened by codelion - 1
Gsm8k bad test
#16 opened by Tostino - 3
Minimal working MCTS example
#5 opened by RomanKoshkin - 2
Too many tokens
#4 opened by integral-llc - 2