Software Open Access
Wolf, Thomas; Debut, Lysandre; Sanh, Victor; Chaumond, Julien; Delangue, Clement; Moi, Anthony; Cistac, Perric; Ma, Clara; Jernite, Yacine; Plu, Julien; Xu, Canwen; Le Scao, Teven; Gugger, Sylvain; Drame, Mariama; Lhoest, Quentin; Rush, Alexander M.
The Nyströmformer model was proposed in Nyströmformer: A Nyström-Based Algorithm for Approximating Self-Attention by Yunyang Xiong, Zhanpeng Zeng, Rudrasis Chakraborty, Mingxing Tan, Glenn Fung, Yin Li, and Vikas Singh.
The Nyströmformer model overcomes the quadratic complexity of self-attention on the input sequence length by adapting the Nyström method to approximate standard self-attention, enabling longer sequences with thousands of tokens as input.
Compatible checkpoints can be found on the hub: https://huggingface.co/models?other=nystromformer
REALMThe REALM model was proposed in REALM: Retrieval-Augmented Language Model Pre-Training by Kelvin Guu, Kenton Lee, Zora Tung, Panupong Pasupat and Ming-Wei Chang.
It's a retrieval-augmented language model that firstly retrieves documents from a textual knowledge corpus and then utilizes retrieved documents to process question answering tasks.
Compatible checkpoints can be found on the hub: https://huggingface.co/models?other=realm
ViTMAEThe ViTMAE model was proposed in Masked Autoencoders Are Scalable Vision Learners by Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, Ross Girshick.
The paper shows that, by pre-training a Vision Transformer (ViT) to reconstruct pixel values for masked patches, one can get results after fine-tuning that outperform supervised pre-training.
Compatible checkpoints can be found on the hub: https://huggingface.co/models?other=vit_mae
ViLTThe ViLT model was proposed in ViLT: Vision-and-Language Transformer Without Convolution or Region Supervision by Wonjae Kim, Bokyung Son, Ildoo Kim.
ViLT incorporates text embeddings into a Vision Transformer (ViT), allowing it to have a minimal design for Vision-and-Language Pre-training (VLP).
Compatible checkpoints can be found on the hub: https://huggingface.co/models?other=vilt
Swin TransformerThe Swin Transformer was proposed in Swin Transformer: Hierarchical Vision Transformer using Shifted Windows by Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, Baining Guo.
The Swin Transformer serves as a general-purpose backbone for computer vision. The shifted windowing scheme brings greater efficiency by limiting self-attention computation to non-overlapping local windows while also allowing for cross-window connection. This hierarchical architecture has the flexibility to model at various scales and has linear computational complexity with respect to image size.
Compatible checkpoints can be found on the hub: https://huggingface.co/models?other=swin
YOSOThe YOSO model was proposed in You Only Sample (Almost) Once: Linear Cost Self-Attention Via Bernoulli Sampling by Zhanpeng Zeng, Yunyang Xiong, Sathya N. Ravi, Shailesh Acharya, Glenn Fung, Vikas Singh.
YOSO approximates standard softmax self-attention via a Bernoulli sampling scheme based on Locality Sensitive Hashing (LSH). In principle, all the Bernoulli random variables can be sampled with a single hash.
Compatible checkpoints can be found on the hub: https://huggingface.co/models?other=yoso
Add model likeTo help contributors add new models more easily to Transformers, there is a new command that will clone an existing model and set the various hooks in the library, so that you only have to write the tweaks needed to the modeling file. Just run transformers-cli add-new-model-like
and fill the questionnaire!
New training scripts were introduced, for speech seq2seq models and an image pre-training script leveraging the ViTMAE models. Finally, an image captioning example in Flax gets added to the library.
Adding support for long files on automatic-speech-recognition
(ASR) as well as supporting audio models with LM which increases the WER on many tasks See the blogpost.
Also continuously increasing homogeneity in arguments, framework support on all pipelines.
TF
on image-classification
pipeline. by @Narsil in https://github.com/huggingface/transformers/pull/15030batch_size
enabled on zero-cls
and qa
pipelines. by @Narsil in https://github.com/huggingface/transformers/pull/14225The ELECTRA model can now be used as a decoder, enabling an ELECTRA encoder-decoder model.
ElectraForCausalLM
-> Enable Electra encoder-decoder model by @stancld in https://github.com/huggingface/transformers/pull/14729The vision encoder decoder model can now be used in TensorFlow.
CLIP gets ported to TensorFlow.
RoFormer gets ported to Flax.
--optim
by @manuelciosici in https://github.com/huggingface/transformers/pull/14744The documentation has been fully migrated to MarkDown, if you are making contribution, make sure to read the upgraded guide on how to write good docstrings.
AttributeError
from PreTrainedTokenizerFast.decoder
by @aphedges in https://github.com/huggingface/transformers/pull/14691run_name
in MLflowCallback by @YangDong2002 in https://github.com/huggingface/transformers/pull/14894num_return_sequences
support for text2text generation. by @Narsil in https://github.com/huggingface/transformers/pull/14988tokenizers
upgrade. by @Narsil in https://github.com/huggingface/transformers/pull/14941chunk_length_s
instead of _ms
. by @Narsil in https://github.com/huggingface/transformers/pull/15029batch_size
arg (like others enabled everywhere). by @Narsil in https://github.com/huggingface/transformers/pull/15027with torch.no_grad()
to DistilBERT integration test forward pass by @jaketae in https://github.com/huggingface/transformers/pull/14979tokenize_chinese_chars
arg by @SaulLu in https://github.com/huggingface/transformers/pull/15158np.ndarray
optional arguments by @gante in https://github.com/huggingface/transformers/pull/15074is_ctc
needs to be updated to `self.type == "ctc". by @Narsil in https://github.com/huggingface/transformers/pull/15194from_encoder_decoder_pretrained
in encoder-decoder models by @jsnfly in https://github.com/huggingface/transformers/pull/15056The community contributors below have significantly contributed to the v4.16.0 release. Thank you!
Full Changelog: https://github.com/huggingface/transformers/compare/v4.15.0...v4.16.0
Name | Size | |
---|---|---|
huggingface/transformers-v4.16.0.zip
md5:c7645f6587b93c480e6bb38e91591fb0 |
10.0 MB | Download |
All versions | This version | |
---|---|---|
Views | 67,224 | 643 |
Downloads | 2,018 | 8 |
Data volume | 18.6 GB | 80.2 MB |
Unique views | 56,507 | 593 |
Unique downloads | 1,237 | 8 |