llama_cpp_for_radxa_dragon_.../examples
2024-02-11 12:44:51 +01:00
..
baby-llama
batched
batched-bench llama : remove LLAMA_MAX_DEVICES and LLAMA_SUPPORTS_GPU_OFFLOAD (#5240) 2024-01-31 17:30:17 +02:00
batched.swift
beam-search
benchmark
convert-llama2c-to-ggml
embedding
export-lora
finetune
gguf
imatrix Adding some imatrix tools (#5302) 2024-02-04 10:39:58 +02:00
infill Remove unused data and add fixes (#5154) 2024-01-27 15:25:55 +01:00
jeopardy
llama-bench refactor : switch to emplace_back to avoid extra object (#5291) 2024-02-03 13:23:37 +02:00
llama.android
llama.swiftui
llava llava : add requirements.txt and update README.md (#5428) 2024-02-09 15:00:59 +02:00
lookahead
lookup lookup: add print for drafting performance (#5450) 2024-02-11 12:44:51 +01:00
main refactor : switch to emplace_back to avoid extra object (#5291) 2024-02-03 13:23:37 +02:00
main-cmake-pkg
parallel
passkey
perplexity refactor : switch to emplace_back to avoid extra object (#5291) 2024-02-03 13:23:37 +02:00
quantize refactor : switch to emplace_back to avoid extra object (#5291) 2024-02-03 13:23:37 +02:00
quantize-stats refactor : switch to emplace_back to avoid extra object (#5291) 2024-02-03 13:23:37 +02:00
save-load-state
server server : add llama2 chat template (#5425) 2024-02-11 12:16:22 +02:00
simple
speculative
sycl [SYCL] update guide of SYCL backend (#5254) 2024-02-02 15:53:27 +08:00
tokenize
train-text-from-scratch
alpaca.sh
base-translate.sh
chat-13B.bat
chat-13B.sh
chat-persistent.sh
chat-vicuna.sh
chat.sh
CMakeLists.txt ggml : add unified SYCL backend for Intel GPUs (#2690) 2024-01-28 17:56:23 +02:00
gpt4all.sh
json-schema-to-grammar.py
llama.vim
llama2-13b.sh
llama2.sh
llm.vim
make-ggml.py
Miku.sh
pydantic-models-to-grammar-examples.py
pydantic_models_to_grammar.py
reason-act.sh
server-llama2-13B.sh