chore(all): replace instances of interface with any (#10067)

Both interface{} and any (which is just an alias for interface{} introduced in Go 1.18) represent the empty interface that all types satisfy.
This commit is contained in:
Bruce MacDonald
2025-04-02 09:44:27 -07:00
committed by GitHub
parent 4e415029b3
commit 9876c9faa4
19 changed files with 58 additions and 58 deletions

View File

@@ -60,7 +60,7 @@ type Model struct {
System string
License []string
Digest string
Options map[string]interface{}
Options map[string]any
Messages []api.Message
Template *template.Template

View File

@@ -72,7 +72,7 @@ var (
errBadTemplate = errors.New("template error")
)
func modelOptions(model *Model, requestOpts map[string]interface{}) (api.Options, error) {
func modelOptions(model *Model, requestOpts map[string]any) (api.Options, error) {
opts := api.DefaultOptions()
if err := opts.FromMap(model.Options); err != nil {
return api.Options{}, err
@@ -826,7 +826,7 @@ func GetModelInfo(req api.ShowRequest) (*api.ShowResponse, error) {
cs := 30
for k, v := range m.Options {
switch val := v.(type) {
case []interface{}:
case []any:
for _, nv := range val {
params = append(params, fmt.Sprintf("%-*s %#v", cs, k, nv))
}
@@ -1336,7 +1336,7 @@ func Serve(ln net.Listener) error {
return nil
}
func waitForStream(c *gin.Context, ch chan interface{}) {
func waitForStream(c *gin.Context, ch chan any) {
c.Header("Content-Type", "application/json")
for resp := range ch {
switch r := resp.(type) {

View File

@@ -38,7 +38,7 @@ type Scheduler struct {
pendingReqCh chan *LlmRequest
finishedReqCh chan *LlmRequest
expiredCh chan *runnerRef
unloadedCh chan interface{}
unloadedCh chan any
loaded map[string]*runnerRef
loadedMu sync.Mutex
@@ -68,7 +68,7 @@ func InitScheduler(ctx context.Context) *Scheduler {
pendingReqCh: make(chan *LlmRequest, maxQueue),
finishedReqCh: make(chan *LlmRequest, maxQueue),
expiredCh: make(chan *runnerRef, maxQueue),
unloadedCh: make(chan interface{}, maxQueue),
unloadedCh: make(chan any, maxQueue),
loaded: make(map[string]*runnerRef),
newServerFn: llm.NewLlamaServer,
getGpuFn: discover.GetGPUInfo,
@@ -618,8 +618,8 @@ func (runner *runnerRef) needsReload(ctx context.Context, req *LlmRequest) bool
// a before and after GPU memory allocation. The returned channel
// will be notified when we're done waiting, or have timed out and should
// proceed anyway
func (runner *runnerRef) waitForVRAMRecovery() chan interface{} {
finished := make(chan interface{}, 1)
func (runner *runnerRef) waitForVRAMRecovery() chan any {
finished := make(chan any, 1)
// CPU or Metal don't need checking, so no waiting required
// windows can page VRAM, only cuda currently can report accurate used vram usage