Issues
- 1
- 1
- 5
- 1
unable to use memory_extreme
#24 opened by egollhofer - 1
LocalGemma2ForCausalLM.from_pretrained("google/gemma-2-27b-it", preset="memory_extreme") can't load model.
#27 opened by webbigdata-jp - 6
authentication in LocalGemma2ForCausalLM with huggingface access token doesn't work
#21 opened by dsingal0 - 1
The first preset table in the README is wrong
#28 opened by phil71x - 1
8bit quantization
#30 opened by paolo-losi - 8
"auto" shouldn't use "exact" not "memory" if there is enough memory to load the entire model
#20 opened by dsingal0 - 3
- 1
API Provide
#19 opened by codeMonkey-shin