From 75dc800d212bdffe7f6ac7e4bbb5cb311e9889fb Mon Sep 17 00:00:00 2001 From: Jhen-Jie Hong Date: Tue, 7 Nov 2023 16:51:27 +0800 Subject: [PATCH] talk-llama : fix n_gpu_layers usage again (#1442) --- examples/talk-llama/talk-llama.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/talk-llama/talk-llama.cpp b/examples/talk-llama/talk-llama.cpp index bc0119a..8c41ef5 100644 --- a/examples/talk-llama/talk-llama.cpp +++ b/examples/talk-llama/talk-llama.cpp @@ -267,7 +267,7 @@ int main(int argc, char ** argv) { auto lmparams = llama_model_default_params(); if (!params.use_gpu) { - lcparams.lmparams = 0; + lmparams.n_gpu_layers = 0; } struct llama_model * model_llama = llama_load_model_from_file(params.model_llama.c_str(), lmparams);