mirror of
https://github.com/dogkeeper886/ollama37.git
synced 2025-12-11 08:17:03 +00:00
subprocess improvements (#524)
* subprocess improvements - increase start-up timeout - when runner fails to start fail rather than timing out - try runners in order rather than choosing 1 runner - embed metal runner in metal dir rather than gpu - refactor logging and error messages * Update llama.go * Update llama.go * simplify by using glob
This commit is contained in:
10
llm/ggml.go
10
llm/ggml.go
@@ -166,15 +166,15 @@ func (c *containerLORA) Decode(r io.Reader) (model, error) {
|
||||
}
|
||||
|
||||
var (
|
||||
ggmlInit sync.Once
|
||||
ggmlRunnerPath string
|
||||
ggmlInit sync.Once
|
||||
ggmlRunners []ModelRunner // a slice of ModelRunners ordered by priority
|
||||
)
|
||||
|
||||
func ggmlRunner() ModelRunner {
|
||||
func ggmlRunner() []ModelRunner {
|
||||
ggmlInit.Do(func() {
|
||||
ggmlRunnerPath = chooseRunner("ggml")
|
||||
ggmlRunners = chooseRunners("ggml")
|
||||
})
|
||||
return ModelRunner{Path: ggmlRunnerPath}
|
||||
return ggmlRunners
|
||||
}
|
||||
|
||||
const (
|
||||
|
||||
Reference in New Issue
Block a user