Skip to content
#

Transformer

A transformer is a deep learning architecture based on self-attention mechanisms, designed to process sequential data in parallel. Transformers are the foundation of modern large language models and are widely used in natural language processing, computer vision, and generative AI.

Here are 6,835 public repositories matching this topic...

annotated_deep_learning_paper_implementations

🧑‍🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠

  • Updated Jan 22, 2026
  • Python

RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.

  • Updated Mar 30, 2026
  • Python

Easy-to-use Speech Toolkit including Self-Supervised Learning model, SOTA/Streaming ASR with punctuation, Streaming TTS with text frontend, Speaker Verification System, End-to-End Speech Translation and Keyword Spotting. Won NAACL2022 Best Demo Award.

  • Updated Apr 2, 2026
  • Python
Followers
167 followers
Website
github.com/topics/transformer
Wikipedia
Wikipedia