Top Transformers repositories on GitHub
Transformer model implementations, training kits, and finetuning tooling.
Ranked by stars across 863 repositories tagged transformers. Refreshed daily.
- 1microsoft/generative-ai-for-beginners★ 110,289 · ⑂ 59,156
21 Lessons, Get Started Building with Generative AI
- ai
- chatgpt
- dall-e
- generativeai
- gpt
- azure
- 2rasbt/LLMs-from-scratch★ 92,069 · ⑂ 14,225
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
- chatgpt
- gpt
- large-language-models
- llm
- python
- pytorch
- 3hiyouga/LlamaFactory★ 70,990 · ⑂ 8,673
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
- fine-tuning
- llama
- llm
- peft
- transformers
- rlhf
- 4labmlai/annotated_deep_learning_paper_implementations★ 66,547 · ⑂ 6,713
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
- deep-learning
- deep-learning-tutorial
- pytorch
- gan
- transformers
- reinforcement-learning
- 5lucidrains/vit-pytorch★ 25,149 · ⑂ 3,495
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
- artificial-intelligence
- attention-mechanism
- transformers
- computer-vision
- image-classification
- 6deepset-ai/haystack★ 25,102 · ⑂ 2,767
Open-source AI orchestration framework for building context-engineered, production-ready LLM applications. Design modular pipelines and agent workflows with explicit control over retrieval, routing, memory, and generation. Built for scalable agents, RAG, multimodal applications, semantic search, and conversational systems.
- nlp
- question-answering
- pytorch
- semantic-search
- information-retrieval
- summarization
- 7amusi/CVPR2026-Papers-with-Code★ 22,520 · ⑂ 2,793
CVPR 2026 论文和开源项目合集
- cvpr
- cvpr2020
- computer-vision
- deep-learning
- machine-learning
- object-detection
- 8huggingface/peft★ 21,074 · ⑂ 2,278
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
- adapter
- diffusion
- llm
- parameter-efficient-learning
- python
- pytorch
- 9stas00/ml-engineering★ 17,870 · ⑂ 1,137
Machine Learning Engineering Open Book
- pytorch
- slurm
- large-language-models
- llm
- machine-learning
- scalability
- 10arc53/DocsGPT★ 17,870 · ⑂ 2,033
Private AI platform for agents, assistants and enterprise search. Built-in Agent Builder, Deep research, Document analysis, Multi-model support, and API connectivity for agents.
- ai
- python
- natural-language-processing
- react
- chatgpt
- docsgpt
- 11NVIDIA/Megatron-LM★ 16,238 · ⑂ 3,915
Ongoing research training transformer models at scale
- large-language-models
- model-para
- transformers
- 12huggingface/transformers.js★ 15,961 · ⑂ 1,136
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
- browser
- javascript
- transformers
- webml
- 13BlinkDL/RWKV-LM★ 14,513 · ⑂ 1,009
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.
- attention-mechanism
- deep-learning
- gpt
- gpt-2
- gpt-3
- language-model
- 14PaddlePaddle/PaddleNLP★ 12,937 · ⑂ 3,044
Easy-to-use and powerful LLM and SLM library with awesome model zoo.
- nlp
- embedding
- bert
- ernie
- paddlenlp
- pretrained-models
- 15neuml/txtai★ 12,471 · ⑂ 808
💡 All-in-one AI framework for semantic search, LLM orchestration and language model workflows
- python
- search
- nlp
- semantic-search
- vector-search
- txtai
- 16NielsRogge/Transformers-Tutorials★ 11,617 · ⑂ 1,716
This repository contains demos I made with the Transformers library by HuggingFace.
- transformers
- pytorch
- bert
- vision-transformer
- layoutlm
- gpt-2
- 17qubvel-org/segmentation_models.pytorch★ 11,532 · ⑂ 1,835
Semantic segmentation models with 500+ pretrained convolutional and transformer-based backbones.
- segmentation
- image-processing
- pspnet
- unet
- unet-pytorch
- pytorch
- 18speechbrain/speechbrain★ 11,517 · ⑂ 1,686
A PyTorch-based Speech Toolkit
- speech-recognition
- speech-toolkit
- speaker-recognition
- speech-to-text
- speech-enhancement
- speech-separation
- 19huggingface/tokenizers★ 10,699 · ⑂ 1,084
💥 Fast State-of-the-Art Tokenizers optimized for Research and Production
- nlp
- natural-language-processing
- natural-language-understanding
- language-model
- transformers
- bert
- 20openvinotoolkit/openvino★ 10,202 · ⑂ 3,200
OpenVINO™ is an open source toolkit for optimizing and deploying AI inference
- inference
- deep-learning
- openvino
- ai
- computer-vision
- diffusion-models
- 21niedev/RTranslator★ 9,844 · ⑂ 894
Open source real-time translation app for Android that runs locally
- translator
- bluetooth-le
- realtime-translator
- android-app
- onnx
- onnxruntime
- 22OpenRLHF/OpenRLHF★ 9,457 · ⑂ 936
An Easy-to-use, Scalable and High-performance Agentic RL Framework based on Ray (PPO & DAPO & REINFORCE++ & VLM & TIS & vLLM & Ray & Async RL)
- transformers
- vllm
- large-language-models
- raylib
- reinforcement-learning-from-human-feedback
- reinforcement-learning
- 23intel/ipex-llm★ 8,801 · ⑂ 1,424
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, DeepSeek, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, vLLM, DeepSpeed, Axolotl, etc.
- pytorch
- llm
- transformers
- gpu
- 24FoundationVision/VAR★ 8,679 · ⑂ 566
[NeurIPS 2024 Best Paper Award][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simple, user-friendly yet state-of-the-art* codebase for autoregressive image generation!
- auto-regressive-model
- diffusion-models
- image-generation
- transformers
- autoregressive-models
- generative-ai
- 25EleutherAI/gpt-neo★ 8,279 · ⑂ 961
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
- language-model
- transformers
- gpt
- gpt-2
- gpt-3
Find engineers shipping Transformers
The list above ranks the most-starred public repositories tagged with the Transformers topic, drawn from the public GitHub graph. Across 863 repositories tagged this way, the maintainers and top contributors are a tight cluster of the people actually building Transformers.
Looking for engineers who’ve worked on Transformers for real, not just listed it on LinkedIn? The fastest path is the contributor list of these repos. Their commits, issues, and READMEs are public proof of depth.
Refolk turns this list into a search. Ask for “maintainers of top Transformers repos who are hiring”, “Transformers engineers in San Francisco”, or “founders shipping Transformers” and Refolk returns a ranked shortlist with sources.
How this list is built
Last refreshed: Thu, 07 May 2026 05:54:47 GMT
Need a list like this for any search?
Refolk runs natural-language searches across GitHub, LinkedIn, and the open web. Try one of these:
Browse other topics
- Top LLM repos
- Top RAG repos
- Top Vector databases repos
- Top Ethereum repos
- Top Game development repos
- Top Computer vision repos
- Top Docker repos
- Top Speech recognition repos
See all repository lists.