Ritual
GitHubGitHub (opens in a new tab)
  • Home
    • Overview
    • Hello, World!
    • Running an ONNX Model
    • Running a Torch Model
    • TGI Inference with Mistral-7B
    • Prompt to NFT
    • Running GPT-4
    • Introduction
    • Regression
    • Classification
    • Neural Networks
    • Reinforcement Learning
    • Embeddings
    • Learn ML with Examples #1
    • FAQs
    • Introduction to Transformers
    • Input Embedding
    • Positional Embedding
    • Attention
    • FFN & Outputs
    • Transformers Resources

Jump to...

  • Papers
  • Lectures
  • Blog Posts
  • Code
Understanding Transformers
Transformers Resources

Resources

Papers

  • Attention is All You Need (opens in a new tab)
  • Formal Algorithms for Transformers (opens in a new tab)

Lectures

  • Sebastian Raschka's Lecture on Transformers (opens in a new tab)
  • Lets build GPT by Andrej Karpathy (opens in a new tab)
  • Transformers - Advanced NLP Lecture (opens in a new tab)

Blog Posts

  • Self Attention from Scratch - Sebastian Raschka (opens in a new tab)
  • Transformer Code Walkthrough - Mark Riedl (opens in a new tab)

Code

  • GPT in 60 Lines of Numpy (opens in a new tab)
  • Transformer from Scratch in Pytorch (opens in a new tab)
FFN & Outputs

MIT 2024 © Nextra.