-
Notifications
You must be signed in to change notification settings - Fork 11.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
convert safetensor adapters into GGUF #6327
Conversation
// Convert writes an Ollama compatible model to the provided io.WriteSeeker based on configurations | ||
// and files it finds in the input path. | ||
// Supported input model formats include safetensors. | ||
// Supported input tokenizers files include tokenizer.json (preferred) and tokenizer.model. | ||
func Convert(fsys fs.FS, ws io.WriteSeeker) error { | ||
func ConvertModel(fsys fs.FS, ws io.WriteSeeker) error { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Convert should remain the same but call convertAdapter()
internally
08dd531
to
e7c08fe
Compare
convert/convert.go
Outdated
KV(llm.KV) llm.KV | ||
Tensors([]Tensor) []llm.Tensor | ||
|
||
tensorName(string) string |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
#5365 changes how tensor names are mapped so be aware
943f3f3
to
da3d6a2
Compare
|
||
return conv.writeFile(ws, conv.KV(baseKV), conv.Tensors(ts)) | ||
} | ||
|
||
// Convert writes an Ollama compatible model to the provided io.WriteSeeker based on configurations |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Comment needs an update if we're going to change this
@@ -11,12 +11,21 @@ import ( | |||
"github.com/ollama/ollama/llm" | |||
) | |||
|
|||
type Parameters struct { | |||
type ModelParameters struct { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
it's implicit (as it is elsewhere in the product) that we're dealing with models already
type ModelParameters struct { | |
type Parameters struct { |
da3d6a2
to
d722cd5
Compare
d722cd5
to
8360dc0
Compare
This change converts a Safetensors based LoRA into GGUF and ties it w/ a base model. Only llama2/llama3/mistral/gemma2 will work initially. You can create the Modelfile to look like:
I'll add in some tests, but wanted to get this out so people could try it out.
Replaces #5524