Deep (Learning) Focus
Subscribe
Sign in
Home
Notes
The Author
Archive
About
Latest
Top
Discussions
AI Agents from First Principles
Understanding AI agents by building upon the most basic concepts of LLMs...
Jun 9
•
Cameron R. Wolfe, Ph.D.
219
Share this post
Deep (Learning) Focus
AI Agents from First Principles
Copy link
Facebook
Email
Notes
More
21
May 2025
A Guide for Debugging LLM Training Data
Data-centric techniques and tools that anyone should use when training an LLM...
May 19
•
Cameron R. Wolfe, Ph.D.
69
Share this post
Deep (Learning) Focus
A Guide for Debugging LLM Training Data
Copy link
Facebook
Email
Notes
More
1
April 2025
Llama 4: The Challenges of Creating a Frontier-Level LLM
The full story behind Llama 4 and Meta's huge pivot in research strategy...
Apr 28
•
Cameron R. Wolfe, Ph.D.
77
Share this post
Deep (Learning) Focus
Llama 4: The Challenges of Creating a Frontier-Level LLM
Copy link
Facebook
Email
Notes
More
2
March 2025
Vision Large Language Models (vLLMs)
Teaching LLMs to understand images and videos in addition to text...
Mar 31
•
Cameron R. Wolfe, Ph.D.
124
Share this post
Deep (Learning) Focus
Vision Large Language Models (vLLMs)
Copy link
Facebook
Email
Notes
More
11
nanoMoE: Mixture-of-Experts (MoE) LLMs from Scratch in PyTorch
An introductory, simple, and functional implementation of MoE LLM pretraining...
Mar 10
•
Cameron R. Wolfe, Ph.D.
135
Share this post
Deep (Learning) Focus
nanoMoE: Mixture-of-Experts (MoE) LLMs from Scratch in PyTorch
Copy link
Facebook
Email
Notes
More
12
February 2025
Demystifying Reasoning Models
Understanding reasoning models and their relation to standard LLMs...
Feb 18
•
Cameron R. Wolfe, Ph.D.
226
Share this post
Deep (Learning) Focus
Demystifying Reasoning Models
Copy link
Facebook
Email
Notes
More
3
January 2025
Mixture-of-Experts (MoE) LLMs
Understanding models like DeepSeek, Grok, and Mixtral from the ground up...
Jan 27
•
Cameron R. Wolfe, Ph.D.
226
Share this post
Deep (Learning) Focus
Mixture-of-Experts (MoE) LLMs
Copy link
Facebook
Email
Notes
More
10
Scaling Laws for LLMs: From GPT-3 to o3
Understanding the current state of LLM scaling and the future of AI research...
Jan 6
•
Cameron R. Wolfe, Ph.D.
126
Share this post
Deep (Learning) Focus
Scaling Laws for LLMs: From GPT-3 to o3
Copy link
Facebook
Email
Notes
More
8
December 2024
Finetuning LLM Judges for Evaluation
The Prometheus suite, JudgeLM, PandaLM, AutoJ, and more...
Dec 2, 2024
•
Cameron R. Wolfe, Ph.D.
79
Share this post
Deep (Learning) Focus
Finetuning LLM Judges for Evaluation
Copy link
Facebook
Email
Notes
More
4
November 2024
Automatic Prompt Optimization
Practical techniques for improving prompt quality without manual effort...
Nov 4, 2024
•
Cameron R. Wolfe, Ph.D.
88
Share this post
Deep (Learning) Focus
Automatic Prompt Optimization
Copy link
Facebook
Email
Notes
More
9
September 2024
Model Merging: A Survey
From modern LLM applications to the early days of machine learning research...
Sep 16, 2024
•
Cameron R. Wolfe, Ph.D.
74
Share this post
Deep (Learning) Focus
Model Merging: A Survey
Copy link
Facebook
Email
Notes
More
8
July 2024
Using LLMs for Evaluation
LLM-as-a-Judge and other scalable additions to human quality ratings...
Jul 22, 2024
•
Cameron R. Wolfe, Ph.D.
108
Share this post
Deep (Learning) Focus
Using LLMs for Evaluation
Copy link
Facebook
Email
Notes
More
14
Share
Copy link
Facebook
Email
Notes
More
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts