About StaffML
8,053 physics-grounded ML systems questions across 79 topics and 4 deployment tracks, at 6 difficulty levels from recall to system design. Backed by a 600-page open textbook. Free, open source, and runs entirely in your browser.
“Every semester, students come to my office hours with the same question: how do I prepare for ML interviews? Not the modeling side — the infrastructure. The compute, the memory, the hardware, the deployment. These interviews expect you to reason about all of it, often with numbers, on the spot. And most people aren't ready.
That's why I built StaffML. I wanted to give students and engineers a way to find out what they really know, and what they still need to learn. The questions come straight from the Machine Learning Systems textbook, but a textbook teaches you concepts one at a time, and quizzes test whether you remember them. Interviews challenge you to connect concepts across the entire system stack. StaffML does just that.
It's free because interview prep is just another form of education, and education works best when it's free and open access, and the world needs more AI engineers.”
What makes StaffML different
Every question traces back to a specific chapter of the Machine Learning Systems textbook. You are learning the 79 concepts a curriculum designer chose, not whatever the internet happened to cough up.
When a question asks about memory bandwidth, the numbers come from actual H100, A100, and Jetson datasheets. The math works on real silicon, not on round numbers that make the answer tidy.
Questions ask you to estimate, diagnose, compare tradeoffs, and architect — the same skills tested in Staff+ interview loops. Nothing asks you to recite a definition.
Type your calculation, then compare it against the model answer. The app tells you whether you are in the right ballpark or off by a factor of ten.
Every question passed a second-pass math check by a separate model. The initial verification pass flagged an 8.3% error rate across the corpus. All flagged errors were corrected.
AI is not magic — it is infrastructure, and infrastructure has laws.
StaffML is part of the Machine Learning Systems curriculum at Harvard University. Every topic links back to its source chapter.
Try a Question
The FP16 Model Footprint
A model has 7 billion parameters. How much VRAM does it occupy just to load the weights in FP16 precision?
Try this question →Who is this for?
Start with L1–L2 recall questions to build your foundation.
Jump to L4–L6+ questions. Try the Mock Interview.
Do the Daily Challenge — 3 questions, 5 minutes, same for everyone.
Browse the question bank to see what the field looks like.
How Questions Are Organized
Every question is tagged by difficulty (6 levels from recall to system design), competency zone (what kind of thinking it tests), and deployment track (Cloud, Edge, Mobile, or TinyML).
Difficulty Levels
“What does GPU HBM stand for?”
“Why is memory bandwidth often more important than FLOPS for inference?”
“Calculate the minimum batch size to saturate an H100's memory bandwidth.”
“Your serving latency spiked 3x after switching FP16→FP8. Why?”
“Design a serving stack for Llama-70B at 10K QPS on a $50K/month budget.”
“Design fault-tolerant training for a 1T param model across 3 data centers.”
Deployment Tracks
How Questions Are Built
StaffML questions are constructed using LLM-assisted generation with structured prompts grounded in the Machine Learning Systems textbook and the MLSysIM physics engine. Every hardware specification traces back to a centralized constants table maintained alongside the textbook.
Every question undergoes independent math verification by a separate model that rechecks all arithmetic and hardware specs. The initial verification pass found an 8.3% error rate across the corpus. All identified errors were corrected.
Found an error? We take correctness seriously. If you spot a wrong number, a broken calculation, or a misleading scenario, open an issue on GitHub. Community verification is how we keep improving.
Open Source
The entire question corpus, taxonomy, and web application are open source. Contributions, feedback, and corrections are welcome.
View on GitHubv0.0.5-dev · built 2026-04-02
