mirror of
https://github.com/dogkeeper886/ollama37.git
synced 2025-12-10 15:57:04 +00:00
Bump llama.cpp to b2474
The release just before ggml-cuda.cu refactoring
This commit is contained in:
@@ -343,6 +343,12 @@ struct llama_server_context
|
||||
|
||||
~llama_server_context()
|
||||
{
|
||||
if (clp_ctx)
|
||||
{
|
||||
LOG_INFO("freeing clip model", {});
|
||||
clip_free(clp_ctx);
|
||||
clp_ctx = nullptr;
|
||||
}
|
||||
if (ctx)
|
||||
{
|
||||
llama_free(ctx);
|
||||
|
||||
Reference in New Issue
Block a user