From 8a203f9fa1b24e010be8f35ebbbd6786293684cb Mon Sep 17 00:00:00 2001 From: Georgi Gerganov Date: Sat, 20 May 2023 10:14:31 +0300 Subject: [PATCH] llama : fix compile warnings in llama_set_state_data() --- llama.cpp | 4 ++-- llama.h | 2 +- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/llama.cpp b/llama.cpp index 6ebe85d0f..68e3bec13 100644 --- a/llama.cpp +++ b/llama.cpp @@ -2618,8 +2618,8 @@ size_t llama_copy_state_data(struct llama_context * ctx, uint8_t * dst) { } // Sets the state reading from the specified source address -size_t llama_set_state_data(struct llama_context * ctx, const uint8_t * src) { - const uint8_t * inp = src; +size_t llama_set_state_data(struct llama_context * ctx, uint8_t * src) { + uint8_t * inp = src; // set rng { diff --git a/llama.h b/llama.h index fd3f21e5f..8623e08ce 100644 --- a/llama.h +++ b/llama.h @@ -138,7 +138,7 @@ extern "C" { // Set the state reading from the specified address // Returns the number of bytes read - LLAMA_API size_t llama_set_state_data(struct llama_context * ctx, const uint8_t * src); + LLAMA_API size_t llama_set_state_data(struct llama_context * ctx, uint8_t * src); // Save/load session file LLAMA_API bool llama_load_session_file(struct llama_context * ctx, const char * path_session, llama_token * tokens_out, size_t n_token_capacity, size_t * n_token_count_out);