| .. |
|
baby-llama
|
build : fix most gcc and clang warnings (#2861)
|
2023-09-01 16:34:50 +03:00 |
|
beam-search
|
build : do not use _GNU_SOURCE gratuitously (#2035)
|
2023-09-08 15:09:21 +03:00 |
|
benchmark
|
|
|
|
convert-llama2c-to-ggml
|
fix some warnings from gcc and clang-tidy (#3038)
|
2023-09-07 13:22:29 -04:00 |
|
embd-input
|
build : do not use _GNU_SOURCE gratuitously (#2035)
|
2023-09-08 15:09:21 +03:00 |
|
embedding
|
examples : make n_ctx warning work again (#3066)
|
2023-09-08 11:43:35 -04:00 |
|
gguf
|
examples : replace fprintf to stdout with printf (#3017)
|
2023-09-05 15:10:27 -04:00 |
|
gptneox-wip
|
fix some warnings from gcc and clang-tidy (#3038)
|
2023-09-07 13:22:29 -04:00 |
|
jeopardy
|
chmod : make scripts executable (#2675)
|
2023-08-23 17:29:09 +03:00 |
|
llama-bench
|
llama-bench : use two tokens in the warmup run for prompt evals (#3059)
|
2023-09-07 15:52:34 +02:00 |
|
main
|
examples : make n_ctx warning work again (#3066)
|
2023-09-08 11:43:35 -04:00 |
|
metal
|
|
|
|
perplexity
|
examples : make n_ctx warning work again (#3066)
|
2023-09-08 11:43:35 -04:00 |
|
quantize
|
fix some warnings from gcc and clang-tidy (#3038)
|
2023-09-07 13:22:29 -04:00 |
|
quantize-stats
|
fix some warnings from gcc and clang-tidy (#3038)
|
2023-09-07 13:22:29 -04:00 |
|
save-load-state
|
fix some warnings from gcc and clang-tidy (#3038)
|
2023-09-07 13:22:29 -04:00 |
|
server
|
fix some warnings from gcc and clang-tidy (#3038)
|
2023-09-07 13:22:29 -04:00 |
|
simple
|
build : do not use _GNU_SOURCE gratuitously (#2035)
|
2023-09-08 15:09:21 +03:00 |
|
speculative
|
build : do not use _GNU_SOURCE gratuitously (#2035)
|
2023-09-08 15:09:21 +03:00 |
|
train-text-from-scratch
|
fix some warnings from gcc and clang-tidy (#3038)
|
2023-09-07 13:22:29 -04:00 |
|
alpaca.sh
|
|
|
|
chat-13B.bat
|
|
|
|
chat-13B.sh
|
|
|
|
chat-persistent.sh
|
|
|
|
chat-vicuna.sh
|
|
|
|
chat.sh
|
main : log file (#2748)
|
2023-08-30 09:29:32 +03:00 |
|
CMakeLists.txt
|
speculative : PoC for speeding-up inference via speculative sampling (#2926)
|
2023-09-03 15:12:08 +03:00 |
|
gpt4all.sh
|
|
|
|
json-schema-to-grammar.py
|
chmod : make scripts executable (#2675)
|
2023-08-23 17:29:09 +03:00 |
|
llama.vim
|
|
|
|
llama2-13b.sh
|
|
|
|
llama2.sh
|
|
|
|
llm.vim
|
llm.vim : stop generation at multiple linebreaks, bind to <F2> (#2879)
|
2023-08-30 09:50:55 +03:00 |
|
make-ggml.py
|
chmod : make scripts executable (#2675)
|
2023-08-23 17:29:09 +03:00 |
|
Miku.sh
|
|
|
|
reason-act.sh
|
chmod : make scripts executable (#2675)
|
2023-08-23 17:29:09 +03:00 |
|
server-llama2-13B.sh
|
chmod : make scripts executable (#2675)
|
2023-08-23 17:29:09 +03:00 |