transformers
https://github.com/huggingface/transformers
Python
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Triage Issues!
When you volunteer to triage issues, you'll receive an email each day with a link to an open issue that needs help in this project. You'll also receive instructions on how to triage issues.
Triage Docs!
Receive a documented method or class from your favorite GitHub repos in your inbox every day. If you're really pro, receive undocumented methods or classes and supercharge your commit history.
Python not yet supported46 Subscribers
View all SubscribersAdd a CodeTriage badge to transformers
Help out
- Issues
- Remove explicit cuda stream in nemotron_h
- refactor: improved the cli server module code organization
- AutoTokenizer.from_pretrained calls model_info() unconditionally in _patch_mistral_regex, breaks HF_HUB_OFFLINE mode
- Fix Mllama torch.compile failure caused by new attention mask logic
- Transformers Qwen3.5 had a bug when set output_hidden_states=True
- Fix Zamba2MambaMixer ignoring use_mamba_kernels=False
- Add cu_seqlens support to OlmoHybridGatedDeltaNet for packed sequences
- AutoModelForSequenceClassification with attn_implementation="flash_attention_3" causes degenerate training (loss increases, model predicts all-one-class)
- Fix Mistral4 tests
- Add AudioFlamingoNext model
- Docs
- Python not yet supported