-
Notifications
You must be signed in to change notification settings - Fork 433
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Timm support #262
Timm support #262
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Overall, this design looks good. We might want a common schema across huggingface and TIMM though. So for hugging-face, our models would then be:
model
hf:
model_name: '<name_here>'
Would that work for you @moinnadeem ?
Timm is all image classification on imagenet so was super straightforward. looking at HF, they have model factories such as So hf may look like: masked language modeling
question answering
and so on. |
Does |
I don't think so. I'm seeing "Bert-base" show up in multiple examples. Though, I'm not sure of the differences between tasks. It may just be different weights or a different classifier head. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM! 2 comments on docstrings and a question on if we should include a **kwargs
argument.
@mosaicml/research-engineering does anyone know if we could have a kwargs
-like argument in yahp? Maybe defining a dict hparam kwargs: dict = hp.optional(default={})
, then using it in a yaml like: kwargs.image_size: 512
?
YAHP supports JSON types, at least to some extent. I don't have any experience using it. |
Yeah, the In other words, there is a bijection between model names and weights. |
Hm, we have something very similar to this at the moment. I would like to make our current version more into a factory soon (see how we have very thin wrappers around the current HF configs). However, we do need to insert some additional variables, such as |
Co-authored-by: Landan Seguin <landanjs@gmail.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good, a few suggested changes. Do you have a w&b run with the TIMM resnet50 to verify convergence?
https://wandb.ai/mosaic-ml/timm-imagenet/reports/Shared-panel-22-01-30-20-01-86--VmlldzoxNTAzMTky resnet50 convergence run to 76.89 accuracy. note on the yaml said |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
@hanlint , @Landanjs