Bump llama.cpp to b2474

The release just before ggml-cuda.cu refactoring
This commit is contained in:
Daniel Hiltgen
2024-03-23 09:54:56 +01:00
parent a5ba0fcf78
commit 43799532c1
5 changed files with 7 additions and 184 deletions

View File

@@ -343,6 +343,12 @@ struct llama_server_context
~llama_server_context()
{
if (clp_ctx)
{
LOG_INFO("freeing clip model", {});
clip_free(clp_ctx);
clp_ctx = nullptr;
}
if (ctx)
{
llama_free(ctx);