|
|
c4628523bc
|
llama automatic unloading and longer start timeout
|
2025-07-29 02:31:39 +02:00 |
|
|
|
071e87ee44
|
disable warmups
|
2025-07-29 02:24:14 +02:00 |
|
|
|
9e17aadb56
|
add gemma3 model
|
2025-07-29 02:22:52 +02:00 |
|
|
|
0fde3108d6
|
move llama models to ssd
|
2025-07-26 17:54:23 +02:00 |
|
|
|
9765f1cf86
|
add gemma3n
|
2025-07-23 23:46:44 +02:00 |
|
|
|
5f3a00b382
|
add qwen3 no thinking
|
2025-07-23 22:56:52 +02:00 |
|
|
|
b379c181f2
|
increase context size
|
2025-07-23 22:06:45 +02:00 |
|
|
|
e1801347f2
|
add qwen3
|
2025-07-23 20:15:37 +02:00 |
|
|
|
d53db88fd2
|
gpu offload in llama.cpp
|
2025-07-23 19:55:48 +02:00 |
|
|
|
18eb912f03
|
llama-swap
|
2025-07-23 00:18:45 +02:00 |
|