llama_cpp_for_radxa_dragon_.../tools
Georgi Gerganov d7d826b3c1
server : support multi-modal context checkpoints (#19849)
* Modify llama-memory-hybrid-iswa.cpp

* Modify llama-memory-recurrent.cpp

* Modify server-common.cpp

* Modify server-common.h

* Modify server-context.cpp

* Modify server-task.h

* Added comment to llama-memory-hybrid-iswa.cpp

* Remove comment from server-context.cpp

* Stylistic fix server-context.cpp

* Fix an issue when seqrm isn't called in server-context.cpp

* cont : alternative impl

* cont : cleanup

* cont : n_tokens -> int64_t

---------

Co-authored-by: timkhronos <timkhronos@gmail.com>
2026-02-25 15:14:27 +02:00
..
batched-bench
cli cli : provide model with text filename (#19783) 2026-02-22 22:33:49 +01:00
completion llama : remove write/read of output ids/logits/embeddings (#18862) 2026-02-23 07:04:30 +01:00
cvector-generator
export-lora
fit-params
gguf-split
imatrix
llama-bench
mtmd model: Add PaddleOCR-VL model support (#18825) 2026-02-19 17:05:25 +01:00
perplexity perplexity: add proper batching (#19661) 2026-02-16 18:44:44 +02:00
quantize quantize : add --dry-run option (#19526) 2026-02-20 09:20:16 +01:00
rpc NetBSD build support (#19589) 2026-02-14 09:47:01 +01:00
server server : support multi-modal context checkpoints (#19849) 2026-02-25 15:14:27 +02:00
tokenize
tts
CMakeLists.txt