llama.cpp/common
Kerfuffle 6e08281e58
Extend llama_kv_cache_seq_rm to allow matching any sequence (#3843)
* Extend llama_kv_cache_seq_rm to allow matichng any sequence

* Replace llama_kv_cache_tokens_rm with llama_kv_cache_clear

Use llama_kv_cache_clear for cache clearing

Change calls to llama_kv_cache_tokens_rm that want to delete by position to use llama_kv_cache_seq_rm functionality
2023-10-29 11:31:40 -06:00
..
CMakeLists.txt common : fix mirostat state when using multiple sequences (#3543) 2023-10-11 22:35:46 +03:00
common.cpp Extend llama_kv_cache_seq_rm to allow matching any sequence (#3843) 2023-10-29 11:31:40 -06:00
common.h sampling : refactor init to use llama_sampling_params (#3696) 2023-10-20 21:07:23 +03:00
console.cpp check C++ code with -Wmissing-declarations (#3184) 2023-09-15 15:38:27 -04:00
console.h gguf : new file format with flexible meta data (beta) (#2398) 2023-08-21 23:07:43 +03:00
grammar-parser.cpp ggml : fix rope + llama minor optimizations (#3560) 2023-10-20 13:02:12 +03:00
grammar-parser.h gguf : new file format with flexible meta data (beta) (#2398) 2023-08-21 23:07:43 +03:00
log.h log : disable pid in log filenames 2023-10-25 10:09:16 +03:00
sampling.cpp llama : add option for greedy sampling with probs (#3813) 2023-10-28 14:23:11 +03:00
sampling.h sampling : refactor init to use llama_sampling_params (#3696) 2023-10-20 21:07:23 +03:00
stb_image.h examples: support LLaVA v1.5 (multimodal model) (#3436) 2023-10-12 18:23:18 +03:00
train.cpp llama : remove token functions with context args in favor of model (#3720) 2023-10-23 22:40:03 +03:00
train.h train : finetune LORA (#2632) 2023-09-28 21:40:11 +03:00