Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

convert safetensor adapters into GGUF #6327

Merged
merged 20 commits into from
Aug 23, 2024
Merged

convert safetensor adapters into GGUF #6327

merged 20 commits into from
Aug 23, 2024

Conversation

pdevine
Copy link
Contributor

@pdevine pdevine commented Aug 12, 2024

This change converts a Safetensors based LoRA into GGUF and ties it w/ a base model. Only llama2/llama3/mistral/gemma2 will work initially. You can create the Modelfile to look like:

FROM llama3
ADAPTER /path/to/my/safetensor/adapter/directory

I'll add in some tests, but wanted to get this out so people could try it out.

Replaces #5524

@jmorganca jmorganca requested review from mxyng and joshyan1 and removed request for mxyng August 12, 2024 21:34
// Convert writes an Ollama compatible model to the provided io.WriteSeeker based on configurations
// and files it finds in the input path.
// Supported input model formats include safetensors.
// Supported input tokenizers files include tokenizer.json (preferred) and tokenizer.model.
func Convert(fsys fs.FS, ws io.WriteSeeker) error {
func ConvertModel(fsys fs.FS, ws io.WriteSeeker) error {
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Convert should remain the same but call convertAdapter() internally

@pdevine pdevine force-pushed the pdevine/gguf-adapters branch from 08dd531 to e7c08fe Compare August 13, 2024 01:13
KV(llm.KV) llm.KV
Tensors([]Tensor) []llm.Tensor

tensorName(string) string
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

#5365 changes how tensor names are mapped so be aware

@pdevine pdevine force-pushed the pdevine/gguf-adapters branch 2 times, most recently from 943f3f3 to da3d6a2 Compare August 16, 2024 01:22

return conv.writeFile(ws, conv.KV(baseKV), conv.Tensors(ts))
}

// Convert writes an Ollama compatible model to the provided io.WriteSeeker based on configurations
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Comment needs an update if we're going to change this

@@ -11,12 +11,21 @@ import (
"github.com/ollama/ollama/llm"
)

type Parameters struct {
type ModelParameters struct {
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

it's implicit (as it is elsewhere in the product) that we're dealing with models already

Suggested change
type ModelParameters struct {
type Parameters struct {

@pdevine pdevine force-pushed the pdevine/gguf-adapters branch from da3d6a2 to d722cd5 Compare August 20, 2024 01:09
@pdevine pdevine force-pushed the pdevine/gguf-adapters branch from d722cd5 to 8360dc0 Compare August 21, 2024 23:07
@pdevine pdevine merged commit 0c819e1 into main Aug 23, 2024
12 checks passed
@pdevine pdevine deleted the pdevine/gguf-adapters branch August 23, 2024 18:29
deep93333 pushed a commit to deep93333/ollama that referenced this pull request Sep 9, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants