mirror of
https://github.com/dogkeeper886/ollama37.git
synced 2025-12-11 00:07:07 +00:00
envconfig: Remove no longer supported max vram var (#10623)
Co-authored-by: Richard Lyons <frob@cloudstaff.com>
This commit is contained in:
@@ -209,8 +209,6 @@ var (
|
|||||||
MaxRunners = Uint("OLLAMA_MAX_LOADED_MODELS", 0)
|
MaxRunners = Uint("OLLAMA_MAX_LOADED_MODELS", 0)
|
||||||
// MaxQueue sets the maximum number of queued requests. MaxQueue can be configured via the OLLAMA_MAX_QUEUE environment variable.
|
// MaxQueue sets the maximum number of queued requests. MaxQueue can be configured via the OLLAMA_MAX_QUEUE environment variable.
|
||||||
MaxQueue = Uint("OLLAMA_MAX_QUEUE", 512)
|
MaxQueue = Uint("OLLAMA_MAX_QUEUE", 512)
|
||||||
// MaxVRAM sets a maximum VRAM override in bytes. MaxVRAM can be configured via the OLLAMA_MAX_VRAM environment variable.
|
|
||||||
MaxVRAM = Uint("OLLAMA_MAX_VRAM", 0)
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func Uint64(key string, defaultValue uint64) func() uint64 {
|
func Uint64(key string, defaultValue uint64) func() uint64 {
|
||||||
|
|||||||
Reference in New Issue
Block a user