Update dependency transformers to v4.41.0 #455
Merged
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR contains the following updates:
==4.40.2
->==4.41.0
Release Notes
huggingface/transformers (transformers)
v4.41.0
: : Phi3, JetMoE, PaliGemma, VideoLlava, Falcon2 and FalconVLMCompare Source
New models
Phi3
The Phi-3 model was proposed in Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone by Microsoft.
TLDR; Phi-3 introduces new ROPE scaling methods, which seems to scale fairly well! A 3b and a
Phi-3-mini is available in two context-length variants—4K and 128K tokens. It is the first model in its class to support a context window of up to 128K tokens, with little impact on quality.
JetMoE
JetMoe-8B is an 8B Mixture-of-Experts (MoE) language model developed by Yikang Shen and MyShell. JetMoe project aims to provide a LLaMA2-level performance and efficient language model with a limited budget. To achieve this goal, JetMoe uses a sparsely activated architecture inspired by the ModuleFormer. Each JetMoe block consists of two MoE layers: Mixture of Attention Heads and Mixture of MLP Experts. Given the input tokens, it activates a subset of its experts to process them. This sparse activation schema enables JetMoe to achieve much better training throughput than similar size dense models. The training throughput of JetMoe-8B is around 100B tokens per day on a cluster of 96 H100 GPUs with a straightforward 3-way pipeline parallelism strategy.
PaliGemma
PaliGemma is a lightweight open vision-language model (VLM) inspired by PaLI-3, and based on open components like the SigLIP vision model and the Gemma language model. PaliGemma takes both images and text as inputs and can answer questions about images with detail and context, meaning that PaliGemma can perform deeper analysis of images and provide useful insights, such as captioning for images and short videos, object detection, and reading text embedded within images.
More than 120 checkpoints are released see the collection here !
VideoLlava
Video-LLaVA exhibits remarkable interactive capabilities between images and videos, despite the absence of image-video pairs in the dataset.
💡 Simple baseline, learning united visual representation by alignment before projection
With the binding of unified visual representations to the language feature space, we enable an LLM to perform visual reasoning capabilities on both images and videos simultaneously.
🔥 High performance, complementary learning with video and image
Extensive experiments demonstrate the complementarity of modalities, showcasing significant superiority when compared to models specifically designed for either images or videos.
Falcon 2 and FalconVLM:
Two new models from TII-UAE! They published a blog-post with more details! Falcon2 introduces parallel mlp, and falcon VLM uses the
Llava
frameworkGGUF
from_pretrained
supportYou can now load most of the GGUF quants directly with transformers'
from_pretrained
to convert it to a classic pytorch model. The API is simple:We plan more closer integrations with llama.cpp / GGML ecosystem in the future, see: https://github.com/huggingface/transformers/issues/27712 for more details
Quantization
New quant methods
In this release we support new quantization methods: HQQ & EETQ contributed by the community. Read more about how to quantize any transformers model using HQQ & EETQ in the dedicated documentation section
dequantize
API for bitsandbytes modelsIn case you want to dequantize models that have been loaded with bitsandbytes, this is now possible through the
dequantize
API (e.g. to merge adapter weights)dequantize
API for bitsandbytes quantized models by @younesbelkada in https://github.com/huggingface/transformers/pull/30806API-wise, you can achieve that with the following:
Generation updates
Gemma
work withtorch.compile
by @ydshieh in https://github.com/huggingface/transformers/pull/30775Sdpa support
BERT
] Add support for sdpa by @hackyon in https://github.com/huggingface/transformers/pull/28802🚨 might be breaking
evaluation_strategy
toeval_strategy
🚨🚨🚨 by @muellerzr in https://github.com/huggingface/transformers/pull/30190Cleanups
Not breaking but important for Llama tokenizers
LlamaTokenizerFast
] Refactor default llama by @ArthurZucker in https://github.com/huggingface/transformers/pull/28881Fixes
Fix missing
prev_ci_results
by @ydshieh in https://github.com/huggingface/transformers/pull/30313Fix: remove
pad token id
in pipeline forward arguments by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30285fix Parameter dtype in audio models by @ylacombe in https://github.com/huggingface/transformers/pull/30310
disable use_cache if using gradient checkpointing by @chenzizhao in https://github.com/huggingface/transformers/pull/30320
Fix test transposing image with EXIF Orientation tag by @albertvillanova in https://github.com/huggingface/transformers/pull/30319
Avoid
jnp
import inutils/generic.py
by @ydshieh in https://github.com/huggingface/transformers/pull/30322Fix
AssertionError
in clip conversion script by @ydshieh in https://github.com/huggingface/transformers/pull/30321[UDOP] Add special tokens to tokenizer by @NielsRogge in https://github.com/huggingface/transformers/pull/29594
Enable multi-device for some models by @jla524 in https://github.com/huggingface/transformers/pull/30207
feat: Upgrade Weights & Biases callback by @parambharat in https://github.com/huggingface/transformers/pull/30135
[Feature Extractors] Fix kwargs to pre-trained by @sanchit-gandhi in https://github.com/huggingface/transformers/pull/30260
Pipeline: fix
pad_token_id
again by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30338[Whisper] Fix slow tests by @sanchit-gandhi in https://github.com/huggingface/transformers/pull/30152
parallel job limit for doctest by @ydshieh in https://github.com/huggingface/transformers/pull/30342
Transformers Metadata by @LysandreJik in https://github.com/huggingface/transformers/pull/30344
Deprecate default chat templates by @Rocketknight1 in https://github.com/huggingface/transformers/pull/30346
Restore casting of masked_spec_embed by @ylacombe in https://github.com/huggingface/transformers/pull/30336
Update unwrap from accelerate by @SunMarc in https://github.com/huggingface/transformers/pull/29933
Do not remove half seq length in generation tests by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30016
Fix config + attn_implementation in AutoModelForCausalLM.from_pretrained by @hiyouga in https://github.com/huggingface/transformers/pull/30299
Add TF swiftformer by @joaocmd in https://github.com/huggingface/transformers/pull/23342
[Grounding DINO] Add resources by @NielsRogge in https://github.com/huggingface/transformers/pull/30232
Nits for model docs by @merveenoyan in https://github.com/huggingface/transformers/pull/29795
Enable multi-device for more models by @jla524 in https://github.com/huggingface/transformers/pull/30379
GenerationConfig: warn if pad token is negative by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30187
Add FSDP config for CPU RAM efficient loading through accelerate by @helloworld1 in https://github.com/huggingface/transformers/pull/30002
Llama
family, fixuse_cache=False
generation by @ArthurZucker in https://github.com/huggingface/transformers/pull/30380Update docstrings for text generation pipeline by @Rocketknight1 in https://github.com/huggingface/transformers/pull/30343
Terminator strings for generate() by @Rocketknight1 in https://github.com/huggingface/transformers/pull/28932
Fix layerwise GaLore optimizer hard to converge with warmup scheduler by @hiyouga in https://github.com/huggingface/transformers/pull/30372
Jamba: fix left-padding test by @gante in https://github.com/huggingface/transformers/pull/30389
Fix DETA save_pretrained by @qubvel in https://github.com/huggingface/transformers/pull/30326
FIX / PEFT: Pass device correctly to peft by @younesbelkada in https://github.com/huggingface/transformers/pull/30397
Add sdpa and fa2 the Wav2vec2 family. by @kamilakesbi in https://github.com/huggingface/transformers/pull/30121
[docs] LLM inference by @stevhliu in https://github.com/huggingface/transformers/pull/29791
show
-rs
to show skip reasons by @ArthurZucker in https://github.com/huggingface/transformers/pull/30318Add inputs embeds in generation by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30269
[Grounding DINO] Add support for cross-attention in GroundingDinoMultiHeadAttention by @EduardoPach in https://github.com/huggingface/transformers/pull/30364
remove redundant logging from longformer by @riklopfer in https://github.com/huggingface/transformers/pull/30365
fix: link to HF repo/tree/revision when a file is missing by @mapmeld in https://github.com/huggingface/transformers/pull/30406
[tests] add
require_torch_sdpa
for test that needs sdpa support by @faaany in https://github.com/huggingface/transformers/pull/30408Jax: scipy version pin by @gante in https://github.com/huggingface/transformers/pull/30402
Fix on "cache position" for assisted generation by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30068
fix for itemsize => element_size() for torch backwards compat by @winglian in https://github.com/huggingface/transformers/pull/30133
Make EosTokenCriteria compatible with mps by @pcuenca in https://github.com/huggingface/transformers/pull/30376
FIX: re-add bnb on docker image by @younesbelkada in https://github.com/huggingface/transformers/pull/30427
Fix LayoutLMv2 init issue and doctest by @ydshieh in https://github.com/huggingface/transformers/pull/30278
Remove old TF port docs by @Rocketknight1 in https://github.com/huggingface/transformers/pull/30426
Rename torch.run to torchrun by @steven-basart in https://github.com/huggingface/transformers/pull/30405
Fix use_cache for xla fsdp by @alanwaketan in https://github.com/huggingface/transformers/pull/30353
[
LlamaTokenizerFast
] Refactor default llama by @ArthurZucker in https://github.com/huggingface/transformers/pull/28881New model PR needs green (slow tests) CI by @ydshieh in https://github.com/huggingface/transformers/pull/30341
Add llama3 by @ArthurZucker in https://github.com/huggingface/transformers/pull/30334
Fix YOLOS image processor resizing by @qubvel in https://github.com/huggingface/transformers/pull/30436
[
Llava
] + CIs fix red cis and llava integration tests by @ArthurZucker in https://github.com/huggingface/transformers/pull/30440[tests] make test device-agnostic by @faaany in https://github.com/huggingface/transformers/pull/30444
fix uncaught init of linear layer in clip's/siglip's for image classification models by @vasqu in https://github.com/huggingface/transformers/pull/30435
fix jamba slow foward for multi-gpu by @SunMarc in https://github.com/huggingface/transformers/pull/30418
[SegGPT] Fix loss calculation by @EduardoPach in https://github.com/huggingface/transformers/pull/30421
Add
paths
filter to avoid the chance of being triggered by @ydshieh in https://github.com/huggingface/transformers/pull/30453Fix wrong indent in
utils/check_if_new_model_added.py
by @ydshieh in https://github.com/huggingface/transformers/pull/30456[
research_project
] Most of the security issues come from this requirement.txt by @ArthurZucker in https://github.com/huggingface/transformers/pull/29977Neuron: When save_safetensor=False, no need to move model to CPU by @jeffhataws in https://github.com/huggingface/transformers/pull/29703
Enable fp16 on CPU by @muellerzr in https://github.com/huggingface/transformers/pull/30459
Non blocking support to torch DL's by @muellerzr in https://github.com/huggingface/transformers/pull/30465
consistent job / pytest report / artifact name correspondence by @ydshieh in https://github.com/huggingface/transformers/pull/30392
Workflow / ENH: Add SSH into our runners workflow by @younesbelkada in https://github.com/huggingface/transformers/pull/30425
FIX / Workflow: Change tailscale trigger condition by @younesbelkada in https://github.com/huggingface/transformers/pull/30471
FIX / Workflow: Fix SSH workflow bug by @younesbelkada in https://github.com/huggingface/transformers/pull/30474
[fix codellama conversion] by @ArthurZucker in https://github.com/huggingface/transformers/pull/30472
Script for finding candidate models for deprecation by @amyeroberts in https://github.com/huggingface/transformers/pull/29686
Fix SigLip classification doctest by @amyeroberts in https://github.com/huggingface/transformers/pull/30475
Don't run fp16 MusicGen tests on CPU by @amyeroberts in https://github.com/huggingface/transformers/pull/30466
Prevent crash with
WandbCallback
with third parties by @tomaarsen in https://github.com/huggingface/transformers/pull/30477Add WSD scheduler by @visheratin in https://github.com/huggingface/transformers/pull/30231
Fix Issue #29817 Video Classification Task Guide Using Undeclared Variables by @manju-rangam in https://github.com/huggingface/transformers/pull/30457
Make accelerate install non-torch dependent by @muellerzr in https://github.com/huggingface/transformers/pull/30463
Introduce Stateful Callbacks by @muellerzr in https://github.com/huggingface/transformers/pull/29666
Fix Llava for 0-embeddings by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30473
Do not use deprecated
SourceFileLoader.load_module()
in dynamic module loading by @XuehaiPan in https://github.com/huggingface/transformers/pull/30370Add sidebar tutorial for chat models by @Rocketknight1 in https://github.com/huggingface/transformers/pull/30401
Quantization:
HfQuantizer
quant method update by @younesbelkada in https://github.com/huggingface/transformers/pull/30484[docs] Spanish translation of pipeline_tutorial.md by @aaronjimv in https://github.com/huggingface/transformers/pull/30252
FEAT: PEFT support for EETQ by @younesbelkada in https://github.com/huggingface/transformers/pull/30449
Fix the
bitsandbytes
error formatting ("Some modules are dispatched on ...") by @kyo-takano in https://github.com/huggingface/transformers/pull/30494Update
dtype_byte_size
to handle torch.float8_e4m3fn/float8_e5m2 types by @mgoin in https://github.com/huggingface/transformers/pull/30488Use the Keras set_random_seed in tests by @Rocketknight1 in https://github.com/huggingface/transformers/pull/30504
Remove skipping logic now that set_epoch exists by @muellerzr in https://github.com/huggingface/transformers/pull/30501
[
DETR
] Remove timm hardcoded logic in modeling files by @amyeroberts in https://github.com/huggingface/transformers/pull/29038[examples] update whisper fine-tuning by @sanchit-gandhi in https://github.com/huggingface/transformers/pull/29938
Fix GroundingDINO, DPR after BERT SDPA update by @amyeroberts in https://github.com/huggingface/transformers/pull/30506
load_image - decode b64encode and encodebytes strings by @amyeroberts in https://github.com/huggingface/transformers/pull/30192
[SegGPT] Fix seggpt image processor by @EduardoPach in https://github.com/huggingface/transformers/pull/29550
Fix link in dbrx.md by @eitanturok in https://github.com/huggingface/transformers/pull/30509
Allow boolean FSDP options in fsdp_config by @helloworld1 in https://github.com/huggingface/transformers/pull/30439
Pass attn_implementation when using AutoXXX.from_config by @amyeroberts in https://github.com/huggingface/transformers/pull/30507
Fix broken link to Transformers notebooks by @clinty in https://github.com/huggingface/transformers/pull/30512
Update runner tag for PR slow CI by @ydshieh in https://github.com/huggingface/transformers/pull/30535
Fix repo. fetch/checkout in PR slow CI job by @ydshieh in https://github.com/huggingface/transformers/pull/30537
Reenable SDPA's FA2 During Training with torch.compile by @warner-benjamin in https://github.com/huggingface/transformers/pull/30442
Include safetensors as part of
_load_best_model
by @muellerzr in https://github.com/huggingface/transformers/pull/30553Pass
use_cache
in kwargs for GPTNeoX by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30538Enable multi-device for more models by @jla524 in https://github.com/huggingface/transformers/pull/30409
Generate: update links on LLM tutorial doc by @gante in https://github.com/huggingface/transformers/pull/30550
DBRX: make fixup by @gante in https://github.com/huggingface/transformers/pull/30578
Fix seq2seq collator padding by @vasqu in https://github.com/huggingface/transformers/pull/30556
BlipModel: get_multimodal_features method by @XavierSpycy in https://github.com/huggingface/transformers/pull/30438
Add chat templating support for KeyDataset in text-generation pipeline by @DarshanDeshpande in https://github.com/huggingface/transformers/pull/30558
Fix generation doctests by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30263
General PR slow CI by @ydshieh in https://github.com/huggingface/transformers/pull/30540
Remove
use_square_size
after loading by @ydshieh in https://github.com/huggingface/transformers/pull/30567Use text config's vocab size in testing models by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30568
Encoder-decoder models: move embedding scale to nn.Module by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30410
Fix Marian model conversion by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30173
Refactor default chat template warnings by @Rocketknight1 in https://github.com/huggingface/transformers/pull/30551
Fix QA example by @Rocketknight1 in https://github.com/huggingface/transformers/pull/30580
remove jax example by @ArthurZucker in https://github.com/huggingface/transformers/pull/30498
Fix canonical model --model_type in examples by @amyeroberts in https://github.com/huggingface/transformers/pull/30480
Gemma: update activation warning by @pcuenca in https://github.com/huggingface/transformers/pull/29995
Bump gitpython from 3.1.32 to 3.1.41 in /examples/research_projects/decision_transformer by @dependabot in https://github.com/huggingface/transformers/pull/30587
Fix image segmentation example - don't reopen image by @amyeroberts in https://github.com/huggingface/transformers/pull/30481
Improve object detection task guideline by @NielsRogge in https://github.com/huggingface/transformers/pull/29967
Generate: remove deprecated public decoding functions and streamline logic 🧼 by @gante in https://github.com/huggingface/transformers/pull/29956
Fix llava half precision and autocast issues by @frasermince in https://github.com/huggingface/transformers/pull/29721
Fix: failing CI after #30568 by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30599
Fix for Neuron by @michaelbenayoun in https://github.com/huggingface/transformers/pull/30259
Fix memory leak with CTC training script on Chinese languages by @lucky-bai in https://github.com/huggingface/transformers/pull/30358
Fix copies for DBRX - neuron fix by @amyeroberts in https://github.com/huggingface/transformers/pull/30610
fix:missing
output_router_logits
in SwitchTransformers by @lausannel in https://github.com/huggingface/transformers/pull/30573Use
contiguous()
in clip checkpoint conversion script by @ydshieh in https://github.com/huggingface/transformers/pull/30613phi3 chat_template does not support system role by @amitportnoy in https://github.com/huggingface/transformers/pull/30606
Docs: fix
generate
-related rendering issues by @gante in https://github.com/huggingface/transformers/pull/30600Docs: add missing
StoppingCriteria
autodocs by @gante in https://github.com/huggingface/transformers/pull/30617Generate: fix
SinkCache
on Llama models by @gante in https://github.com/huggingface/transformers/pull/30581Fix FX tracing issues for Llama by @michaelbenayoun in https://github.com/huggingface/transformers/pull/30619
Output
None
as attention when layer is skipped by @jonghwanhyeon in https://github.com/huggingface/transformers/pull/30597Fix CI after #30410 by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30612
add mlp bias for llama models by @mayank31398 in https://github.com/huggingface/transformers/pull/30031
Fix W&B run name by @qubvel in https://github.com/huggingface/transformers/pull/30462
HQQ: PEFT support for HQQ by @younesbelkada in https://github.com/huggingface/transformers/pull/30632
Prevent
TextGenerationPipeline._sanitize_parameters
from overriding previously provided parameters by @yting27 in https://github.com/huggingface/transformers/pull/30362Avoid duplication in PR slow CI model list by @ydshieh in https://github.com/huggingface/transformers/pull/30634
[
CI update
] Try to use dockers and no cache by @ArthurZucker in https://github.com/huggingface/transformers/pull/29202Check if the current compiled version of pytorch supports MPS by @jiaqianjing in https://github.com/huggingface/transformers/pull/30664
Hotfix-change-ci by @ArthurZucker in https://github.com/huggingface/transformers/pull/30669
Quantization / HQQ: Fix HQQ tests on our runner by @younesbelkada in https://github.com/huggingface/transformers/pull/30668
Fix llava next tie_word_embeddings config by @SunMarc in https://github.com/huggingface/transformers/pull/30640
Trainer._load_from_checkpoint - support loading multiple Peft adapters by @claralp in https://github.com/huggingface/transformers/pull/30505
Trainer - add cache clearing and the option for batched eval metrics computation by @FoamoftheSea in https://github.com/huggingface/transformers/pull/28769
Fix typo: llama3.md by @mimbres in https://github.com/huggingface/transformers/pull/30653
Respect
resume_download
deprecation by @Wauplin in https://github.com/huggingface/transformers/pull/30620top-k instead of top-p in MixtralConfig docstring by @sorgfresser in https://github.com/huggingface/transformers/pull/30687
Bump jinja2 from 3.1.3 to 3.1.4 in /examples/research_projects/decision_transformer by @dependabot in https://github.com/huggingface/transformers/pull/30680
Bump werkzeug from 3.0.1 to 3.0.3 in /examples/research_projects/decision_transformer by @dependabot in https://github.com/huggingface/transformers/pull/30679
Adding _tie_weights() to prediction heads to support low_cpu_mem_usage=True by @hackyon in https://github.com/huggingface/transformers/pull/29024
Fix
cache_position
initialisation for generation withuse_cache=False
by @nurlanov-zh in https://github.com/huggingface/transformers/pull/30485Word-level timestamps broken for short-form audio by @kamilakesbi in https://github.com/huggingface/transformers/pull/30325
Updated docs of
forward
inIdefics2ForConditionalGeneration
with correctignore_index
value by @zafstojano in https://github.com/huggingface/transformers/pull/30678Bump tqdm from 4.63.0 to 4.66.3 in /examples/research_projects/decision_transformer by @dependabot in https://github.com/huggingface/transformers/pull/30646
Bump tqdm from 4.48.2 to 4.66.3 in /examples/research_projects/visual_bert by @dependabot in https://github.com/huggingface/transformers/pull/30645
Reboot Agents by @aymeric-roucher in https://github.com/huggingface/transformers/pull/30387
Bump tqdm from 4.48.2 to 4.66.3 in /examples/research_projects/lxmert by @dependabot in https://github.com/huggingface/transformers/pull/30644
Separate tokenizer tests by @ArthurZucker in https://github.com/huggingface/transformers/pull/30675
Update
workflow_id
inutils/get_previous_daily_ci.py
by @ydshieh in https://github.com/huggingface/transformers/pull/30695Rename artifact name
prev_ci_results
toci_results
by @ydshieh in https://github.com/huggingface/transformers/pull/30697Add safetensors to model not found error msg for default use_safetensors value by @davidgxue in https://github.com/huggingface/transformers/pull/30602
Pin deepspeed by @muellerzr in https://github.com/huggingface/transformers/pull/30701
Patch CLIP image preprocessor by @rootonchair in https://github.com/huggingface/transformers/pull/30698
Add examples for detection models finetuning by @qubvel in https://github.com/huggingface/transformers/pull/30422
[BitsandBytes] Verify if GPU is available by @NielsRogge in https://github.com/huggingface/transformers/pull/30533
Llava: remove dummy labels by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30706
Add installation of examples requirements in CI by @qubvel in https://github.com/huggingface/transformers/pull/30708
Update object detection guide by @qubvel in https://github.com/huggingface/transformers/pull/30683
Immutability for data collators by @vasqu in https://github.com/huggingface/transformers/pull/30603
Cache: models return input cache type by @gante in https://github.com/huggingface/transformers/pull/30716
Add dynamic resolution input/interpolate position embedding to SigLIP by @davidgxue in https://github.com/huggingface/transformers/pull/30719
Enable dynamic resolution for vivit by @jla524 in https://github.com/huggingface/transformers/pull/30630
Removal of deprecated maps by @LysandreJik in https://github.com/huggingface/transformers/pull/30576
Generate: add
min_p
sampling by @gante in https://github.com/huggingface/transformers/pull/30639Fix image post-processing for OWLv2 by @jla524 in https://github.com/huggingface/transformers/pull/30686
KV cache is no longer a model attribute by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30730
Generate: consistently handle special tokens as tensors by @gante in https://github.com/huggingface/transformers/pull/30624
Update CodeLlama references by @osanseviero in https://github.com/huggingface/transformers/pull/30218
[docs] Update es/pipeline_tutorial.md by @aaronjimv in https://github.com/huggingface/transformers/pull/30684
Update llama3.md, fix typo by @mimbres in https://github.com/huggingface/transformers/pull/30739
mlp_only_layers is more flexible than decoder_sparse_step by @eigen2017 in https://github.com/huggingface/transformers/pull/30552
PEFT / Trainer: Make use of
model.active_adapters()
instead of deprecatedmodel.active_adapter
whenever possible by @younesbelkada in https://github.com/huggingface/transformers/pull/30738[docs] Update link in es/pipeline_webserver.md by @aaronjimv in https://github.com/huggingface/transformers/pull/30745
hqq - fix weight check in check_quantized_param by @mobicham in https://github.com/huggingface/transformers/pull/30748
[awq] replace scale when we have GELU by @SunMarc in https://github.com/huggingface/transformers/pull/30074
Workflow: Replace
actions/post-slack
with centrally defined workflow by @younesbelkada in https://github.com/huggingface/transformers/pull/30737Blip dynamic input resolution by @zafstojano in https://github.com/huggingface/transformers/pull/30722
[GroundingDino] Adding ms_deform_attn kernels by @EduardoPach in https://github.com/huggingface/transformers/pull/30768
Llama: fix custom 4D masks, v2 by @poedator in https://github.com/huggingface/transformers/pull/30348
Generation / FIX: Fix multi-device generation by @younesbelkada in https://github.com/huggingface/transformers/pull/30746
Qwen: incorrect setup flag by @gante in https://github.com/huggingface/transformers/pull/30776
enable Pipeline to get device from model by @faaany in https://github.com/huggingface/transformers/pull/30534
[Object detection pipeline] Lower threshold by @NielsRogge in https://github.com/huggingface/transformers/pull/30710
Generate: remove near-duplicate sample/greedy copy by @gante in https://github.com/huggingface/transformers/pull/30773
Port IDEFICS to tensorflow by @a8nova in https://github.com/huggingface/transformers/pull/26870
Generate: assistant should be greedy in assisted decoding by @gante in https://github.com/huggingface/transformers/pull/30778
Save other CI jobs' result (torch/tf pipeline, example, deepspeed etc) by @ydshieh in https://github.com/huggingface/transformers/pull/30699
Deprecate models script by @amyeroberts in https://github.com/huggingface/transformers/pull/30184
skip low_cpu_mem_usage tests by @SunMarc in https://github.com/huggingface/transformers/pull/30782
CI: update to ROCm 6.0.2 and test MI300 by @fxmarty in https://github.com/huggingface/transformers/pull/30266
Fix OWLv2 Doc by @jla524 in https://github.com/huggingface/transformers/pull/30794
Fix cache type in Idefics2 by @zucchini-nlp in https://github.com/huggingface/transformers/pull/30729
PEFT: Access active_adapters as a property in Trainer by @pashminacameron in https://github.com/huggingface/transformers/pull/30790
CI: more models wo cache support by @gante in https://github.com/huggingface/transformers/pull/30780
Deprecate TF weight conversion since we have full Safetensors support now by @Rocketknight1 in https://github.com/huggingface/transformers/pull/30786
[T5] Adding
model_parallel = False
toT5ForTokenClassification
andMT5ForTokenClassification
by @retarfi in https://github.com/huggingface/transformers/pull/30763Added the necessay import of module by @ankur0904 in https://github.com/huggingface/transformers/pull/30804
Configuration
📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined).
🚦 Automerge: Enabled.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
This PR has been generated by Mend Renovate. View repository job log here.