Navigate:
All Repostransformers
~$TRANSF0.1%

Transformers: Model framework for NLP and vision

Unified API for pre-trained transformer models across frameworks.

LIVE RANKINGS • 06:52 AM • STEADY
TOP 100TOP 100TOP 100TOP 100TOP 100TOP 100TOP 100TOP 100TOP 100TOP 100TOP 100TOP 100
OVERALL
#57
13
AI & ML
#31
5
30 DAY RANKING TREND
ovr#57
·AI#31
STARS
154.8K
FORKS
31.7K
DOWNLOADS
122.0K
7D STARS
+157
7D FORKS
+23
Tags:
See Repo:
Share:

Learn more about transformers

Transformers is a Python library that centralizes model definitions for transformer architectures and related deep learning models. It implements model classes compatible with PyTorch and other frameworks, allowing the same model definition to work across different training and inference backends. The library includes pre-trained model weights, tokenizers, and feature extractors for various tasks including natural language processing, computer vision, speech recognition, and multimodal applications. Models defined in Transformers integrate with training frameworks like DeepSpeed, FSDP, and PyTorch-Lightning, as well as inference engines like vLLM.


1

Framework-Agnostic Models

Model definitions are decoupled from specific frameworks, enabling the same model to be used across PyTorch, TensorFlow, and JAX backends. Training and inference tools build on a shared model specification without reimplementation.

2

Unified Model Hub

Direct integration with a centralized repository containing thousands of pre-trained checkpoints. Single function call loads models with weights, configurations, and tokenizers, eliminating manual download and setup.

3

Multi-Modal Architecture Support

Provides unified interfaces for text, vision, audio, and multi-modal models. Process different data types with consistent APIs, enabling straightforward development of applications combining multiple modalities.


from transformers import pipeline

classifier = pipeline("sentiment-analysis")
result = classifier("I love using transformers!")
print(result)


vv4.57.1

Patch release fixing an optional dependency parsing error with optax that broke poetry installations.

  • Update poetry lockfiles if you encountered optax dependency resolution errors in v4.57.0.
  • Verify FSDP workflows with pre-quantized BitsAndBytes checkpoints now load correctly after internal fix.
vv4.57.0

Adds five new model architectures (Qwen3-Next, Qwen3-VL, VaultGemma, LongCat-Flash, FlexOlmo) with no breaking changes reported.

  • Integrate Qwen3-Next for 80B-parameter MoE with 1:50 sparsity, delivering 10x throughput over 32K context vs. Qwen3-32B.
  • Adopt VaultGemma (1B params) for differential-privacy workloads with ε≤2.0 sequence-level DP guarantee via DP-SGD training.
vv4.56.2

Patch release fixing processor multi-processing, RoPE in Jetmoe, config getter regression, and Emu3 dtype parsing.

  • Update if you use processors with multi-processing or rely on Jetmoe RoPE correctness.
  • Fixes config attribute getter regression and Emu3 dtype edge case; no migration required.

See how people are using transformers

Loading tweets...


[ EXPLORE MORE ]

Related Repositories

Discover similar tools and frameworks used by developers