Files
ollama37/convert/convert_llama.go
Michael Yang 55760195e6 fix mllama conversion (#10716)
cross attention Q and K projections needs to have their heads swapped, similar to non-cross attention Q and K tensors
2025-05-15 12:15:01 -07:00

6.5 KiB