Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Where to download pre-trained model weights? #10

Open
SaharIravani opened this issue Jan 29, 2025 · 4 comments
Open

Where to download pre-trained model weights? #10

SaharIravani opened this issue Jan 29, 2025 · 4 comments

Comments

@SaharIravani
Copy link

SaharIravani commented Jan 29, 2025

Hello,

Thanks for the great work.
I am trying to use Orthrus and follow the setup instructions. However, I cannot find any information about where to download the pre-trained model weights.

  • Are pre-trained models available for download or will they be?
  • If so, could you provide a link or instructions on how to obtain them?
  • If not, is training from scratch the only option?

Any guidance would be greatly appreciated.

Thank you!

@phil-fradkin
Copy link
Collaborator

Hello thanks for your interest in our work! The weights are available on huggingface

The pre-trained Orthrus models are available at https://huggingface.co/antichronology/orthrus (you will find it in files and versions), with individual models available for inference at: https://huggingface.co/quietflamingo/orthrus-base-4-track and https://huggingface.co/quietflamingo/orthrus-large-6-track

These should be easier to perform inference on and would look something like:

from transformer import AutoModel

model = AutoModel.from_pretrained("quietflamingo/orthrus-large-6-track")

Please let us know if you have any other issues!

Also all the embeddings are pre-computed and are available for download from the zenood repository
https://zenodo.org/records/14708163

@wconnell
Copy link

wconnell commented Jan 31, 2025

Hi there, is it possible to share a minimal working example of successfully loading the HF model and performing inference? Preferably in a notebook format with required installations?

I am attempting to work through this on a google colab GPU, but the specific dependencies are causing issues (ie mamba-ssm).

@SaharIravani
Copy link
Author

@phil-fradkin, Thank you very much for providing this information. I'll give a try and get back to you as soon as I successfully run it.

@IanShi1996
Copy link
Collaborator

IanShi1996 commented Feb 2, 2025

Hi @wconnell, here's a link to a Colab that should work out-of-the-box: https://colab.research.google.com/drive/1Rb6VC92YoKRPyF2LG4m8zIXjDszm1NZW?usp=sharing

We've also updated the HF model cards to show a (hopefully working) example of Orthrus embedding:
https://huggingface.co/quietflamingo/orthrus-large-6-track

We'll double check the dependencies specified in the requirements file. Since mamba-ssm is under active dev, the newer versions don't always support the PyTorch and CUDA versions we used to pre-train the model. I believe the following set of core dependencies should work:

PyTorch 2.2.2
CUDA 12.1
causal_conv1d==1.2.0.post2
mamba-ssm==1.2.0.post1

Let me know if you run into any further issues!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants