llama_cpp_for_radxa_dragon_.../tests
Georgi Gerganov 2f74c354c0
graph : make FA compatible with MLA + add initial Metal kernels (#12953)
* graph : make mla compatible with FA

* metal : add exp FA kernels for DeepSeek models

ggml-ci

* llama : minor naming updates

ggml-ci

* ggml : disable FA for DS head sizes

* tests : add FA tests for MLA shapes

ggml-ci
2025-04-17 18:16:36 +03:00
..
.gitignore
CMakeLists.txt cmake : enable curl by default (#12761) 2025-04-07 13:35:19 +02:00
get-model.cpp
get-model.h
run-json-schema-to-grammar.mjs
test-arg-parser.cpp common : refactor downloading system, handle mmproj with -hf option (#12694) 2025-04-01 23:44:05 +02:00
test-autorelease.cpp
test-backend-ops.cpp graph : make FA compatible with MLA + add initial Metal kernels (#12953) 2025-04-17 18:16:36 +03:00
test-barrier.cpp
test-c.c
test-chat-template.cpp ci: detach common from the library (#12827) 2025-04-09 10:11:11 +02:00
test-chat.cpp tool-call: fix non-tool-calling grammar crashes w/ Qwen / Hermes 2 templates (#12900) 2025-04-11 21:47:52 +02:00
test-double-float.cpp
test-gguf.cpp
test-grammar-integration.cpp
test-grammar-llguidance.cpp upgrade to llguidance 0.7.10 (#12576) 2025-03-26 11:06:09 -07:00
test-grammar-parser.cpp
test-json-schema-to-grammar.cpp tool-call: fix Qwen 2.5 Coder support, add micro benchmarks, support trigger patterns for lazy grammars (#12034) 2025-03-05 13:05:13 +00:00
test-llama-grammar.cpp
test-log.cpp
test-lora-conversion-inference.sh
test-model-load-cancel.cpp
test-opt.cpp
test-quantize-fns.cpp tests : fix test-quantize-fns to init the CPU backend (#12306) 2025-03-10 14:07:15 +02:00
test-quantize-perf.cpp
test-rope.cpp
test-sampling.cpp
test-tokenizer-0.cpp
test-tokenizer-0.py
test-tokenizer-0.sh
test-tokenizer-1-bpe.cpp
test-tokenizer-1-spm.cpp
test-tokenizer-random.py