llama_cpp_for_radxa_dragon_.../common
Aman Gupta ab14019821
Support diffusion models: Add Dream 7B (#14644)
* Support diffusion models: Add Dream 7B

* Move diffusion to examples

* Move stuff to examples. Add patch to not use kv-cache

* Address review comments

* Make sampling fast

* llama: remove diffusion functions

* Add basic timings + cleanup

* More cleanup

* Review comments: better formating, use LOG instead std::cerr, re-use batch, use ubatch instead of max_length

* fixup!

* Review: move everything to diffusion-cli for now
2025-07-16 20:03:51 +08:00
..
arg.cpp Support diffusion models: Add Dream 7B (#14644) 2025-07-16 20:03:51 +08:00
arg.h
base64.hpp
build-info.cpp.in
chat-parser.cpp llama-chat : Do not throw when tool parsing fails (#14012) 2025-06-14 17:25:15 +01:00
chat-parser.h llama-chat : Do not throw when tool parsing fails (#14012) 2025-06-14 17:25:15 +01:00
chat.cpp server : support jinja extra template kwargs (Qwen3 enable_thinking feature), from command line and from client (#13196) 2025-06-29 20:02:53 +02:00
chat.h server : support jinja extra template kwargs (Qwen3 enable_thinking feature), from command line and from client (#13196) 2025-06-29 20:02:53 +02:00
CMakeLists.txt cmake : do not search for curl libraries by ourselves (#14613) 2025-07-10 15:29:05 +03:00
common.cpp server : pre-calculate EOG logit biases (#14721) 2025-07-16 14:04:12 +03:00
common.h Support diffusion models: Add Dream 7B (#14644) 2025-07-16 20:03:51 +08:00
console.cpp
console.h
json-partial.cpp
json-partial.h
json-schema-to-grammar.cpp common : use std::string_view now that we target c++17 (#14319) 2025-06-22 08:37:43 +03:00
json-schema-to-grammar.h
llguidance.cpp
log.cpp
log.h
ngram-cache.cpp
ngram-cache.h
regex-partial.cpp
regex-partial.h
sampling.cpp
sampling.h
speculative.cpp
speculative.h