-
Notifications
You must be signed in to change notification settings - Fork 28.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
AudioClassificationPipelineTests::test_small_model_pt_fp16 fails for CUDA/XPU but passes for CPU #36340
Comments
Hello, |
It's indeed failing on our T4 runners. Thank you for pointing this out 👍 |
Changes: * Benchmarking scripts are pruned from Transformers by v4.49.0 due to deprecation. So we don't need to test them anymore. * Some cuda specific tests were generalized to cover non-cuda devices which uncovered some issues. * Some new tests were added which fail for both cuda and xpu. * Few regressions due to changes on Transformers side Fixed tests: * huggingface/transformers@b912f5e * `tests/models/git/test_modeling_git.py::GitModelTest::test_inputs_embeds_matches_input_ids` * huggingface/transformers@b5aaf87 * `tests/pipelines/test_pipelines_video_classification.py::VideoClassificationPipelineTests::test_small_model_pt` * `tests/test_pipeline_mixin.py::VideoClassificationPipelineTests::test_small_model_pt` * huggingface/transformers@42c8ccf * `tests/generation/test_utils.py::GenerationIntegrationTests::test_generated_length_assisted_generation` * huggingface/transformers@9fd123a * `test_model_parallelization` * `test_model_parallel_equal_results` Commits which added new tests (or enabled previously skipped tests) which fail: * huggingface/transformers@23d782e * `tests/pipelines/test_pipelines_text_generation.py::TextGenerationPipelineTests::test_return_dict_in_generate` * `tests/test_pipeline_mixin.py::TextGenerationPipelineTests::test_return_dict_in_generate` * huggingface/transformers@2fa876d * `test_cpu_offload` (some of) * `test_disk_offload_bin` (some of) * `test_disk_offload_safetensors` (some of) * `tests/pipelines/test_pipelines_text_generation.py::TextGenerationPipelineTests::test_small_model_pt_bloom_accelerate` * huggingface/transformers@be2ac09 * `tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationModelTest::test_generate_compilation_all_outputs` * `tests/models/paligemma2/test_modeling_paligemma2.py::PaliGemma2ForConditionalGenerationModelTest::test_generate_compilation_all_outputs` * huggingface/transformers#36340 * `tests/pipelines/test_pipelines_audio_classification.py::AudioClassificationPipelineTests::test_small_model_pt_fp16` * huggingface/transformers@1fae54c * `tests/trainer/test_trainer.py::TrainerIntegrationPrerunTest::test_gradient_accumulation_loss_alignment_with_model_loss` * huggingface/transformers@15ec971 * `tests/models/qwen2_5_vl/test_processor_qwen2_5_vl.py::Qwen2_5_VLProcessorTest::test_chat_template_video_custom_sampling` * `tests/models/qwen2_5_vl/test_processor_qwen2_5_vl.py::Qwen2_5_VLProcessorTest::test_chat_template_video_special_processing` Regressions: * huggingface/transformers@365fecb * `tests/generation/test_utils.py::GenerationIntegrationTests::test_encoder_decoder_generate_attention_mask` * huggingface/transformers@da334bc * `tests/generation/test_utils.py::GenerationIntegrationTests::test_generate_input_features_as_encoder_kwarg` * huggingface/transformers@bcfc9d7 * `tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_config` * huggingface/transformers#36267 * `tests/utils/test_import_utils.py` * huggingface/transformers#36267 * `tests/models/marian/test_modeling_marian.py` Signed-off-by: Dmitry Rogozhkin <dmitry.v.rogozhkin@intel.com>
With:
On:
The commit f19135a introduced
AudioClassificationPipelineTests::test_small_model_pt_fp16
test which passes on CPU, but fails if running on CUDA or XPU. PR was:Logs (for cuda):
Note: for XPU (upstream pytorch XPU, not IPEX) log is actually the same including same scores for labels
Does CUDA/XPU work correctly in this test? (I am confused seeing same scores
0.0833
for all 4 labels).Overall, expectation is that test passes for CUDA/XPU or excluded for pytorch device backends if it's CPU specific.
CC: @jiqing-feng @ydshieh
The text was updated successfully, but these errors were encountered: