diff --git a/ggml-cuda.cu b/ggml-cuda.cu index 50df20edd..0b12a9e76 100644 --- a/ggml-cuda.cu +++ b/ggml-cuda.cu @@ -2835,7 +2835,7 @@ void ggml_cuda_transform_tensor(void * data, struct ggml_tensor * tensor) { } void ggml_cuda_free_data(struct ggml_tensor * tensor) { - if (tensor->backend != GGML_BACKEND_GPU && tensor->backend != GGML_BACKEND_GPU_SPLIT) { + if (!tensor || (tensor->backend != GGML_BACKEND_GPU && tensor->backend != GGML_BACKEND_GPU_SPLIT) ) { return; } diff --git a/llama.cpp b/llama.cpp index f48a6ca79..7419b03b6 100644 --- a/llama.cpp +++ b/llama.cpp @@ -194,8 +194,8 @@ struct llama_layer { }; struct llama_kv_cache { - struct ggml_tensor * k; - struct ggml_tensor * v; + struct ggml_tensor * k = NULL; + struct ggml_tensor * v = NULL; struct ggml_context * ctx = NULL;