- GenAI
- ComputerVision
- LLM
- Transformers
- YOLO
- NLP
- external-services
•
•
•
•
•
•
-
Encoder vs Decoder: Understanding BERT, GPT and Modern LLM Architectures
A deep dive into encoder-only, decoder-only, and encoder-decoder architectures, and how models like BERT, GPT, and BART differ.
-
A Deep Dive into Attention: Self-Attention, Multi-Head Attention and Positional Encoding
A comprehensive guide to attention mechanisms in Transformers, including intuition, QKV, self-attention, multi-head attention, and positional encoding.
-
Transformer Architecture Explained: Attention is All You Need
A deep dive into Transformer architecture, including encoder-decoder structure, attention mechanism, positional encoding, and multi-head attention.
-
How Do LLMs Work? Understanding Next Token Prediction
A simple but deep explanation of how Large Language Models work using next-token prediction and probability.
-
What is Generative AI? Types and Architectures
An overview of Generative AI, including text, image, and audio generation, and the architectures behind them.