* llama : add inference support and model types for T5 and FLAN-T5 model families * llama : add new API functions to support encoder-decoder models: llama_encode(), llama_model_has_encoder(), llama_model_decoder_start_token() * common, llama-cli, llama-batched : add support for encoder-decoder models * convert-hf : handle shared token embeddings tensors in T5Model * convert-hf : add support for SentencePiece BPE tokenizer in T5Model (for Pile-T5 models) * convert-hf : add MT5ForConditionalGeneration and UMT5ForConditionalGeneration to architectures supported by T5Model * convert : add t5 tokenizer tests, use "slow" HF tokenizer for t5 --------- Co-authored-by: Stanisław Szymczyk <sszymczy@gmail.com> Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> |
||
|---|---|---|
| .. | ||
| baby-llama | ||
| batched | ||
| batched-bench | ||
| batched.swift | ||
| benchmark | ||
| convert-llama2c-to-ggml | ||
| cvector-generator | ||
| embedding | ||
| eval-callback | ||
| export-lora | ||
| finetune | ||
| gbnf-validator | ||
| gguf | ||
| gguf-split | ||
| gritlm | ||
| imatrix | ||
| infill | ||
| jeopardy | ||
| llama-bench | ||
| llama.android | ||
| llama.swiftui | ||
| llava | ||
| lookahead | ||
| lookup | ||
| main | ||
| main-cmake-pkg | ||
| parallel | ||
| passkey | ||
| perplexity | ||
| quantize | ||
| quantize-stats | ||
| retrieval | ||
| rpc | ||
| save-load-state | ||
| server | ||
| simple | ||
| speculative | ||
| sycl | ||
| tokenize | ||
| train-text-from-scratch | ||
| base-translate.sh | ||
| chat-13B.bat | ||
| chat-13B.sh | ||
| chat-persistent.sh | ||
| chat-vicuna.sh | ||
| chat.sh | ||
| CMakeLists.txt | ||
| convert-legacy-llama.py | ||
| json-schema-pydantic-example.py | ||
| json_schema_to_grammar.py | ||
| llama.vim | ||
| llm.vim | ||
| Miku.sh | ||
| pydantic-models-to-grammar-examples.py | ||
| pydantic_models_to_grammar.py | ||
| reason-act.sh | ||
| regex-to-grammar.py | ||
| server-embd.py | ||
| server-llama2-13B.sh | ||
| ts-type-to-grammar.sh | ||