Hands-on Large Language Models
I am thrilled to introduce the Hands-On Large Language Models book I wrote with Jay Alammar!
With the incredible pace of LLM development, learning about these techniques can be overwhelming. Throughout this book, we take an intuition first approach through visual storytelling with almost 300 custom-made images in the final release.
This book is for those interested in this exciting field. Whether you are a beginner or more advanced, we believe there is something to be found for everyone!
All of the code is freely available on Github making it easy for you to get started with the inner working of LLMs.
Visual Storytelling
Over the course of our careers, Jay and I have created educative content in AI with a large focus on illustrations. With the complex and often mathematics-heavy nature of AI, the field quickly becomes overwhelming. By leveraging a visual style of diving into these techniques, we aim to focus on understanding these techniques moreso than writing down the equations that make them up.
Great examples of our visual styles can be found here:
- The Illustrated Transformer - Jay Alammar
- A Visual Guide to Mamba and State Space Models - Maarten Grootendorst
If these styles appeal to you, then you will definitely like this book!
Table of Contents
This is the current Table of Content which might change in the upcoming months:
-
Part 1 - Theory
- Introduction to Language Models
- Token Embeddings
- Looking Inside Transformer LLMs Part 2 - Using Pre-Trained Language Models
- Text Classification
- Text Clustering and Topic Modeling
- Prompt Engineering
- Advanced Text Generation TEchniques and Tools
- Semantic Search and Retrieval Augmented Generation
- Multimodal Large Language Models Part 3 - Training and Fine-Tuning
- Creating Text Embedding Models
- Fine-Tuning Representation Models for Classification
- Fine-Tuning Generation Models