-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add inference API of AMR #40
Conversation
The test has failed... I'm checking. |
setup.py
Outdated
@@ -5,6 +5,6 @@ | |||
version='0.1', | |||
install_requires=['easydict', 'pandas', 'tqdm', 'pyyaml', 'scikit-learn', 'ffmpeg-python', | |||
'ftfy', 'regex', 'einops', 'fvcore', 'gradio', 'torchlibrosa', 'librosa', | |||
'clip@git+https://github.com/openai/CLIP.git'], | |||
'clip@git+https://github.com/openai/CLIP.git', 'msclap'], |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could you add msclap before 'clip@git+https://github.com/openai/CLIP.git' ?
training/evaluate.py
Outdated
@@ -439,7 +439,7 @@ def check_valid_combination(dataset, feature, domain): | |||
is_valid = check_valid_combination(args.dataset, args.feature, args.domain) | |||
|
|||
if is_valid: | |||
option_manager = BaseOptions(args.model, args.dataset, args.feature, args.domain) | |||
option_manager = BaseOptions(args.model, args.dataset, args.feature, False, args.domain) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could you remove magic number of False? Instead, please insert variable X = False and then use it here for readability?
@h-munakata In addition, no tests are newly provided. Hence, could you add some tests for CLAP feature extractor and AMR inference API? |
I see. I will add some tests in |
For the test and the Gradio demo, I think {moment | qd | cg | ...}-detr for CLAP features is useful. |
Yes, I agree. Could you prepare the pre-trained weights on your Zenodo? |
Sure. I will upload after the training. |
@h-munakata Sorry, I misunderstood your question. I think that in your paper, CG-DETR (or QD-DETR) achieved the highest performance. So, no need for training. All you need to do is upload the current trained models on Zenodo and accessible from Lighthouse. |
My intention in training all models is to add a CLAP feature to the double for loop of FEATURE and MODEL used in the demo and test to make it easier to handle.
I see. I'll upload the model in the next commit for the test. |
@h-munakata BTW, could you finish implementing web demo tomorrow? I will tag the current version as v1.0, and wondering whether you can finish this implementation by tomorrow. |
Yes, I want to make it in time for the DCASE workshop the day after tomorrow. |
9edff28
to
e0b4aca
Compare
Summary
lighthouse/models.py
I implemented
encode_audio()
separately fromencode_video()
.Class AudioEncoder
inlighthouse/feature_extractor/audio_encoder.py
only have model selector andlighthouse/feature_extractor/audio_encoders/{clap_a | pann}.py
has individual model.Future work