llama.cpp/common
slaren 2833a6f63c
ggml-cuda : fix f16 mul mat (#3961)
* ggml-cuda : fix f16 mul mat

ggml-ci

* silence common.cpp warning (bonus)
2023-11-05 18:45:16 +01:00
..
build-info.cpp.in build : link against build info instead of compiling against it (#3879) 2023-11-02 08:50:16 +02:00
CMakeLists.txt cmake : fix relative path to git submodule index (#3915) 2023-11-02 21:40:31 +02:00
common.cpp ggml-cuda : fix f16 mul mat (#3961) 2023-11-05 18:45:16 +01:00
common.h speculative : change default p_accept to 0.5 + CLI args (#3919) 2023-11-03 09:41:56 +02:00
console.cpp check C++ code with -Wmissing-declarations (#3184) 2023-09-15 15:38:27 -04:00
console.h gguf : new file format with flexible meta data (beta) (#2398) 2023-08-21 23:07:43 +03:00
grammar-parser.cpp ggml : fix rope + llama minor optimizations (#3560) 2023-10-20 13:02:12 +03:00
grammar-parser.h gguf : new file format with flexible meta data (beta) (#2398) 2023-08-21 23:07:43 +03:00
log.h log : make generating separate log files optional (#3787) 2023-11-01 16:18:27 +02:00
sampling.cpp sampling : null grammar field after reset (#3885) 2023-11-01 15:40:43 +02:00
sampling.h samplers : Min-P sampler implementation [alternative to Top P/Top K] (#3841) 2023-10-31 20:44:49 +01:00
stb_image.h examples: support LLaVA v1.5 (multimodal model) (#3436) 2023-10-12 18:23:18 +03:00
train.cpp finetune : add -ngl parameter (#3762) 2023-11-01 13:49:04 +02:00
train.h finetune : add -ngl parameter (#3762) 2023-11-01 13:49:04 +02:00