Back to Vault

About StaffML

8,053 physics-grounded ML systems questions across 79 topics and 4 deployment tracks, at 6 difficulty levels from recall to system design. Backed by a 600-page open textbook. Free, open source, and runs entirely in your browser.

No accountsNo tracking100% freeOpen source
8,053
questions
79
topics
4
tracks
6
levels

“Every semester, students come to my office hours with the same question: how do I prepare for ML interviews? Not the modeling side — the infrastructure. The compute, the memory, the hardware, the deployment. These interviews expect you to reason about all of it, often with numbers, on the spot. And most people aren't ready.

That's why I built StaffML. I wanted to give students and engineers a way to find out what they really know, and what they still need to learn. The questions come straight from the Machine Learning Systems textbook, but a textbook teaches you concepts one at a time, and quizzes test whether you remember them. Interviews challenge you to connect concepts across the entire system stack. StaffML does just that.

It's free because interview prep is just another form of education, and education works best when it's free and open access, and the world needs more AI engineers.”

VR
Vijay Janapa ReddiProfessor, Harvard University

What makes StaffML different

Textbook-grounded, not scraped

Every question traces back to a specific chapter of the Machine Learning Systems textbook. You are learning the 79 concepts a curriculum designer chose, not whatever the internet happened to cough up.

Real hardware, real numbers

When a question asks about memory bandwidth, the numbers come from actual H100, A100, and Jetson datasheets. The math works on real silicon, not on round numbers that make the answer tidy.

Systems reasoning, not trivia

Questions ask you to estimate, diagnose, compare tradeoffs, and architect — the same skills tested in Staff+ interview loops. Nothing asks you to recite a definition.

Napkin math with feedback

Type your calculation, then compare it against the model answer. The app tells you whether you are in the right ballpark or off by a factor of ten.

Independently verified

Every question passed a second-pass math check by a separate model. The initial verification pass flagged an 8.3% error rate across the corpus. All flagged errors were corrected.

AI is not magic — it is infrastructure, and infrastructure has laws.

StaffML is part of the Machine Learning Systems curriculum at Harvard University. Every topic links back to its source chapter.

Try a Question

L2memoryglobal

The FP16 Model Footprint

A model has 7 billion parameters. How much VRAM does it occupy just to load the weights in FP16 precision?

Try this question →

Who is this for?

Preparing for your first ML role?

Start with L1–L2 recall questions to build your foundation.

Start Easy
Working engineer targeting Staff+?

Jump to L4–L6+ questions. Try the Mock Interview.

Mock Interview
Short on time?

Do the Daily Challenge — 3 questions, 5 minutes, same for everyone.

Daily Challenge
Just curious about ML systems?

Browse the question bank to see what the field looks like.

Browse

How Questions Are Organized

Every question is tagged by difficulty (6 levels from recall to system design), competency zone (what kind of thinking it tests), and deployment track (Cloud, Edge, Mobile, or TinyML).

Difficulty Levels

L1
Recall(Entry)

What does GPU HBM stand for?

L2
Understand(Junior)

Why is memory bandwidth often more important than FLOPS for inference?

L3
Apply(Mid-Level)

Calculate the minimum batch size to saturate an H100's memory bandwidth.

L4
Analyze(Senior)

Your serving latency spiked 3x after switching FP16→FP8. Why?

L5
Evaluate(Staff)

Design a serving stack for Llama-70B at 10K QPS on a $50K/month budget.

L6+
Architect(Principal)

Design fault-tolerant training for a 1T param model across 3 data centers.

Deployment Tracks

CloudGPU clusters, large-scale training and serving
EdgeOn-device inference, real-time constraints
MobilePhones, power and thermal budgets
TinyMLMicrocontrollers, sub-milliwatt inference

How Questions Are Built

StaffML questions are constructed using LLM-assisted generation with structured prompts grounded in the Machine Learning Systems textbook and the MLSysIM physics engine. Every hardware specification traces back to a centralized constants table maintained alongside the textbook.

Every question undergoes independent math verification by a separate model that rechecks all arithmetic and hardware specs. The initial verification pass found an 8.3% error rate across the corpus. All identified errors were corrected.

Read the Research PaperThe full methodology: backward design from textbook chapters, four-axis taxonomy, LLM-assisted generation pipeline, independent math verification, and the ikigai-inspired competency zone framework.

Found an error? We take correctness seriously. If you spot a wrong number, a broken calculation, or a misleading scenario, open an issue on GitHub. Community verification is how we keep improving.

Open Source

The entire question corpus, taxonomy, and web application are open source. Contributions, feedback, and corrections are welcome.

View on GitHub

v0.0.5-dev · built 2026-04-02