add llama.cpp to librechat
This commit is contained in:
@@ -59,6 +59,19 @@ spec:
|
|||||||
summaryModel: "current_model"
|
summaryModel: "current_model"
|
||||||
forcePrompt: false
|
forcePrompt: false
|
||||||
modelDisplayLabel: "Ollama"
|
modelDisplayLabel: "Ollama"
|
||||||
|
- name: "Llama.cpp"
|
||||||
|
apiKey: "llama"
|
||||||
|
baseURL: "http://llama.llama.svc.cluster.local:11434/v1/chat/completions"
|
||||||
|
models:
|
||||||
|
default: [
|
||||||
|
"DeepSeek-R1-0528-Qwen3-8B-GGUF"
|
||||||
|
]
|
||||||
|
titleConvo: true
|
||||||
|
titleModel: "current_model"
|
||||||
|
summarize: false
|
||||||
|
summaryModel: "current_model"
|
||||||
|
forcePrompt: false
|
||||||
|
modelDisplayLabel: "Llama.cpp"
|
||||||
imageVolume:
|
imageVolume:
|
||||||
enabled: true
|
enabled: true
|
||||||
size: 10G
|
size: 10G
|
||||||
|
|||||||
Reference in New Issue
Block a user