|
|
3e59786c83
|
manually update llama-swap image tag
|
2026-03-05 19:27:45 +01:00 |
|
|
|
96a09ae6f9
|
Merge pull request 'Update caddy Docker tag to v2.11.1' (#141) from renovate/caddy-2.x into fresh-start
Reviewed-on: #141
|
2026-03-02 17:26:21 +00:00 |
|
|
|
5c4535beb6
|
Add mmproj-url for Qwen3.5-35B-A3B-heretic model
|
2026-03-02 03:19:16 +01:00 |
|
|
|
44aa0c8136
|
add gemma-3-270m-it-qat model
|
2026-02-28 23:20:13 +01:00 |
|
|
|
902004f2e7
|
Add Qwen3.5-35B-A3B-heretic models
|
2026-02-28 18:33:42 +01:00 |
|
|
|
bf1f1c0b41
|
Add always loaded Qwen3-VL-2B-Instruct
|
2026-02-28 17:48:20 +01:00 |
|
|
|
5915b8dd30
|
Add Qwen3.5-35-A3B model
|
2026-02-28 15:49:59 +01:00 |
|
|
|
c14257842a
|
automatically fit models by llama.cpp
|
2026-02-26 01:38:39 +01:00 |
|
|
|
d053342234
|
fix models mount
|
2026-02-26 01:25:21 +01:00 |
|
|
|
2dbd964c28
|
add schema reference to config.yaml
|
2026-02-26 00:43:16 +01:00 |
|
|
|
7712aac0f5
|
configure llama-swap to log llama.cpp output
|
2026-02-26 00:39:58 +01:00 |
|
|
|
c7bc79f574
|
add Qwen3-Coder-Next model
|
2026-02-26 00:10:53 +01:00 |
|
|
|
6cba277b9d
|
update llama-swap image
|
2026-02-25 19:07:10 +01:00 |
|
|
|
bfb089aeff
|
migrate llama models to ssd
|
2026-02-25 16:03:12 +01:00 |
|
|
|
ed83a66a83
|
add ssd volume for llama models
|
2026-02-25 15:43:42 +01:00 |
|
|
|
b4ba66dc18
|
Update caddy Docker tag to v2.11.1
|
2026-02-24 00:00:41 +00:00 |
|
|
|
b95c9e7c69
|
switch llama models dir to lvm hdd
|
2026-02-21 16:51:04 +01:00 |
|
|
|
05c28d0d46
|
add lvm hdd llama models pvc
|
2026-02-21 16:28:06 +01:00 |
|
|
|
b21f8e402b
|
add abliterated versions of qwen3-vl
|
2025-12-06 23:33:56 +01:00 |
|
|
|
65e75a4d39
|
Add 8B and 2B variants of qwen3-vl
|
2025-11-15 22:21:10 +01:00 |
|
|
|
6c7457d095
|
fix Qwen3-VL-4B-Instruct-GGUF models looping issue
|
2025-11-15 20:40:27 +01:00 |
|
|
|
9b556e98a9
|
add qwen3-vl thinking variant
|
2025-11-15 19:31:53 +01:00 |
|
|
|
202ebc7b86
|
add qwen3-vl, fix librechat taking over settings and clean up llama config
|
2025-11-15 19:18:43 +01:00 |
|
|
|
ec61023f74
|
fix cache location after llama-swap update
|
2025-11-15 18:05:12 +01:00 |
|
|
|
05d3493bb7
|
update llama-swap
|
2025-11-15 17:57:46 +01:00 |
|
|
|
f4a865ce7a
|
update llama-swap docker image
|
2025-10-19 20:38:39 +02:00 |
|
|
|
c0f9670837
|
Update caddy Docker tag to v2.10.2
|
2025-10-19 18:18:35 +00:00 |
|
|
|
708ffe203c
|
Add Qwen2.5-VL models
|
2025-09-13 02:42:21 +02:00 |
|
|
|
9c61d47fda
|
add qwen3-4b-2507 model
|
2025-08-18 02:50:46 +02:00 |
|
|
|
444c4faf96
|
move all ingresses to new nginx ingress
|
2025-08-03 18:17:37 +02:00 |
|
|
|
a26a351396
|
update llama-swap
|
2025-08-03 17:16:25 +02:00 |
|
|
|
c4628523bc
|
llama automatic unloading and longer start timeout
|
2025-07-29 02:31:39 +02:00 |
|
|
|
071e87ee44
|
disable warmups
|
2025-07-29 02:24:14 +02:00 |
|
|
|
9e17aadb56
|
add gemma3 model
|
2025-07-29 02:22:52 +02:00 |
|
|
|
0fde3108d6
|
move llama models to ssd
|
2025-07-26 17:54:23 +02:00 |
|
|
|
9765f1cf86
|
add gemma3n
|
2025-07-23 23:46:44 +02:00 |
|
|
|
5f3a00b382
|
add qwen3 no thinking
|
2025-07-23 22:56:52 +02:00 |
|
|
|
b379c181f2
|
increase context size
|
2025-07-23 22:06:45 +02:00 |
|
|
|
e1801347f2
|
add qwen3
|
2025-07-23 20:15:37 +02:00 |
|
|
|
d53db88fd2
|
gpu offload in llama.cpp
|
2025-07-23 19:55:48 +02:00 |
|
|
|
18eb912f03
|
llama-swap
|
2025-07-23 00:18:45 +02:00 |
|