mirror of
https://github.com/dogkeeper886/ollama37.git
synced 2025-12-11 08:17:03 +00:00
enable cache_prompt by default
This commit is contained in:
@@ -234,8 +234,8 @@ func predict(llm extServer, opts api.Options, ctx context.Context, predict Predi
|
||||
"penalize_nl": opts.PenalizeNewline,
|
||||
"seed": opts.Seed,
|
||||
"stop": opts.Stop,
|
||||
"image_data": imageData,
|
||||
"cache_prompt": opts.Cache,
|
||||
"image_data": imageData,
|
||||
"cache_prompt": true,
|
||||
}
|
||||
|
||||
if predict.Format == "json" {
|
||||
|
||||
Reference in New Issue
Block a user