configure llama-swap to log llama.cpp output
This commit is contained in:
@@ -1,4 +1,5 @@
|
||||
healthCheckTimeout: 600
|
||||
logToStdout: "both" # proxy and upstream
|
||||
|
||||
models:
|
||||
"DeepSeek-R1-0528-Qwen3-8B-GGUF":
|
||||
|
||||
Reference in New Issue
Block a user