Support for Microsoft Phi-2 model
fishfree opened this issue · 4 comments
⚠️ ⚠️ ⚠️ ⚠️ ⚠️
Hi! I'm a bot running with LocalAI ( a crazy experiment of @mudler ) - please beware that I might hallucinate sometimes!
but.... I can also be funny or helpful 😸 and I can provide generally speaking good tips or places where to look after in the documentation or in the code based on what you wrote in the issue.
Don't engage in conversation with me,
I don't support (yet) replying!
⚠️ ⚠️ ⚠️ ⚠️ ⚠️
I'm not sure what the issue is with the link you provided, but it seems to be related to a model called "phi-2" from the Microsoft Cognitive Services API. If you could provide more context or information about the issue, I can try to help further.
Sources:
this should actually just work with current LocalAI: https://huggingface.co/TheBloke/phi-2-GGUF
This works here with current master and this YAML config (note GPU settings):
name: phi-2
context_size: 2048
f16: true
threads: 11
gpu_layers: 90
mmap: true
parameters:
model: huggingface://TheBloke/phi-2-GGUF/phi-2.Q8_0.gguf
temperature: 0.2
top_k: 40
top_p: 0.95
template:
chat: &template |
Instruct: {{.Input}}
Output:
completion: *template
keeping it open just to have the example added here https://github.com/mudler/LocalAI/tree/master/examples/configurations