llama.cpp/scripts
Georgi Gerganov 4760e7cc0b
sync : ggml (backend v2) (#3912)
* sync : ggml (backend v2) (wip)

* sync : migrate examples and llama.cpp to dynamic graphs (wip)

* sync : update tests + fix max op params to 64

ggml-ci

* sync : ggml-cuda

ggml-ci

* llama : fix save/load state context size

ggml-ci

* sync : try to fix build on tvOS

* sync : pass custom graph sizes in training examples

* sync : update graph copies to new ggml API

* sync : update sync-ggml.sh with new files

* scripts : fix header in sync script

* train : fix context size calculations

* llama : increase inference graph size up to 4096 nodes

* train : allocate grads for backward graphs

* train : allocate grads for gb_tmp
2023-11-13 14:16:23 +02:00
..
build-info.cmake build : link against build info instead of compiling against it (#3879) 2023-11-02 08:50:16 +02:00
build-info.sh build : link against build info instead of compiling against it (#3879) 2023-11-02 08:50:16 +02:00
convert-gg.sh scripts : helper convert script 2023-08-27 15:24:58 +03:00
get-wikitext-2.sh chmod : make scripts executable (#2675) 2023-08-23 17:29:09 +03:00
LlamaConfig.cmake.in cmake : fix transient definitions in find pkg (#3411) 2023-10-02 12:51:49 +03:00
qnt-all.sh scripts : add pipefail 2023-08-29 10:50:30 +03:00
run-all-perf.sh scripts : add pipefail 2023-08-29 10:50:30 +03:00
run-all-ppl.sh scripts : add pipefail 2023-08-29 10:50:30 +03:00
server-llm.sh scripts : add server-llm.sh (#3868) 2023-11-01 11:29:07 +02:00
sync-ggml.sh sync : ggml (backend v2) (#3912) 2023-11-13 14:16:23 +02:00
verify-checksum-models.py scripts : use /usr/bin/env in shebang (#3313) 2023-09-22 23:52:23 -04:00