25 Machine Learning Interview Questions for 2026 (And How Senior Candidates Actually Answer Them)
Preparing for an ML interview in 2026? Review 25 machine learning interview questions with detailed answers covering algorithms, bias-variance, overfitting, and real-world model thinking.
I've spent the better part of fifteen years placing machine learning engineers and data scientists at companies ranging from Series B startups to hyperscalers. I've sat across the table from hundreds of hiring managers and watched them evaluate candidates. And the single most consistent thing I've noticed is this: the people who get offers aren't necessarily the ones with the most technical knowledge. They're the ones who can demonstrate that their knowledge translates to judgment.
A junior candidate can recite the bias-variance tradeoff. A senior candidate can tell you which direction they'd pull first, why, and what they'd look at to know if it worked. That difference, from definition to decision is what every question on this list is really testing.
The 25 questions below were selected by cross-referencing what's actually being asked in interview loops right now. I focused on the topics that show up repeatedly, and specifically on the ones where I consistently see candidates stumble not because they don't know the concept, but because they've rehearsed the textbook definition and stopped there. The sample answers are written the way I coach strong candidates: direct, honest about trade-offs, and light on jargon for its own sake.
One more thing before we get into it: these questions span fundamentals, evaluation, production ML, and system design. That's intentional. Modern ML interview loops, especially at mid-to-senior level, mix all of these. If you're only drilling one category, you're leaving something on the table.
Part 1: The Fundamentals (And Why They're Harder Than They Look)
ere's something I've noticed after years of debriefs: the fundamentals round tells me more about a candidate than any other. Not because the questions are hard - they're not. Because the answers reveal whether someone has worked with these concepts or just read about them.
1. Bias vs. Variance: Explain the difference and what you'd do if each is too high
What the interviewer is really testing:
Can you connect a theoretical concept to a practical diagnostic process? Strong candidates don't just explain the tradeoff - they describe how they'd identify which problem they have and what they'd actually change.
Sample answer:
I think of bias as the model being too rigidm it's missing real signal and underfitting. Variance is the opposite: the model is too reactive, memorizing quirks in the training data and overfitting. In practice, I diagnose it with learning curves and a clean validation strategy.
If I see high bias, I'll add signal, better features, a less constrained model family, or relaxed regularization. If I see high variance, I'll simplify or stabilize: stronger regularization, fewer leaky or high-cardinality features, more data if it's genuinely available. The part I'm careful about is not trying to "tune my way out" of what's actually a data problem. Sometimes the fix is upstream, inconsistent labeling, train/serve skew, or the wrong target variable entirely.
Follow-up to expect:
"Walk me through a real example." / "What would you look at first, the curves, residuals, or something else?"
2. Overfitting vs. Underfitting: How do you detect each early and respond?
What the interviewer is really testing:
Beyond knowing the definitions, can you describe an experimentation discipline? The best candidates have a methodical response, not a list of fixes they've heard about.
Sample answer:
Overfitting is when the model performs well on training data but degrades on unseen data. Underfitting is when it's bad everywhere - it can't represent the underlying pattern. My default prevention approach is: get the split right first, establish a baseline, and only then add complexity.
For overfitting specifically: regularization, early stopping for iterative models, more robust features, and a double-check for leakage. For underfitting: expand features or move to a model that can capture nonlinearities. The mistake I see repeatedly is treating "add more layers" as the universal solution. Sometimes you're just amplifying noise.
Follow-up to expect:
"What would you look for in the learning curves to distinguish the two?"
3. Supervised vs. Unsupervised vs. Reinforcement Learning: What is the Practical Difference?
What the interviewer is really testing:
Not the textbook definitionsm everyone knows those. They want to know whether you think about problem framing before jumping to an approach.
Sample answer:
If I have a clear target label and I'm predicting it, that's supervised. If I'm trying to discover structure, clusters, embeddings, segments without labels, that's unsupervised. Reinforcement learning is different in kind: I'm optimizing a policy based on delayed rewards, with explicit exploration/exploitation trade-offs.
In interviews, I like to ground this in practical constraints. If I'm ranking content and can measure reward online, RL is theoretically possible but I'd still start with supervised learning-to-rank before jumping there. RL has a much higher operational overhead, and in most production settings, that overhead isn't justified unless you've exhausted supervised approaches.
Follow-up to expect:
"When would you not use ML at all?". This is a great senior signal question.
Part 2: Evaluation (Where Most Candidates Leave Points on the Table)
Evaluation questions are where I've seen the biggest gap between strong and weak candidates. Everyone knows what precision and recall are. Very few people can articulate when one matters more than the other, and even fewer connect metric selection to business costs. That gap is what you're trying to close in this section.
4. Cross-Validation: What it is and how you choose the right split
What the interviewer is really testing:
Statistical rigor and awareness of the ways a naive split can produce misleading results.
Sample answer:
Cross-validation estimates generalization by repeatedly training on subsets and validating on held-out folds. The key is choosing a split that actually reflects your production environment. For i.i.d. tabular data, k-fold is usually fine. For time series, I avoid random folds entirely and use time-aware splits. For user-level data, I split by user to prevent leakage across sessions.
I also decide early whether I'm trying to optimize model selection or estimate final performance, those require slightly different practices, and conflating them is a common source of overly optimistic evaluation results.
Follow-up to expect:
"How would you handle this with grouped data?" / "How does your approach change when labels are scarce?"
5. Confusion Matrix: Walk me through one and explain what you do with it
What the interviewer is really testing:
Whether you can go from a diagnostic tool to a decision, not just describe what the four cells mean.
Sample answer:
A confusion matrix counts true positives, false positives, true negatives, and false negatives. I treat it as the source of truth behind most classification metrics. Practically, it tells me what kind of errors I'm making and whether I should adjust the decision threshold, rebalance the data, or rethink the model entirely.
In binary classification, I'm watching which cell dominates. If false negatives are costly - say, missed fraud or missed diagnoses, I'll tune toward higher recall. If false positives are costly - say, false alarms that erode user trust, I'll tune toward precision. The right answer depends entirely on the cost structure of the problem.
Follow-up to expect:
"How would you pick a threshold?" / "How does this change when you have more than two classes?"
Machine learning glossary quick brush-up
Scan this in two minutes before your interview. These are the terms candidates most often stumble on.
Model assumptions too rigid.
Model reacts too strongly to noise.
Great on training data, weak on new data.
Model too simple for the data.
Controls model complexity.
Optimization that reduces model error.
How many predicted positives were correct.
How many real positives you captured.
Balance of precision and recall.
How well classes separate across thresholds.
Better metric for rare positives.
Breakdown of prediction outcomes.
Training data contains future information.
Input distribution changes.
Relationship between inputs and outputs changes.
One label dominates the dataset.
Filling missing values.
Categories replaced with outcome statistics.
Independent models combined.
Sequential models fixing errors.
Multiple train/test splits.
Dimensionality reduction.
Centralized feature management.
Training and production data mismatch.
6. Precision vs. Recall vs. F1: When do you use each?
What the interviewer is really testing:
Business alignment. They want to know you can choose a metric that reflects real-world costs, not just one that looks good.
Sample answer:
Precision is "when I say positive, how often am I right." Recall is "of all the actual positives out there, how many did I catch." F1 balances them as a harmonic mean, which is useful when classes are imbalanced and I need a single number to optimize. But I'm cautious about defaulting to F1 without thinking through the cost structure.
In fraud detection, for example, I'll often bias toward recall early, catch more fraud, even if that means more false positives and then build review workflows or second-stage models to manage the precision problem. The metric I choose should reflect the cost of each error type in the real deployment environment, not just what produces the highest number on a leaderboard.
Follow-up to expect:
"Give me an example where accuracy is a misleading metric." (Imbalanced classes, almost always the right answer here.)
7. ROC-AUC vs. PR-AUC: When does ROC-AUC lie to you?
What the interviewer is really testing:
This separates people who've used metrics from people who understand them.
Sample answer:
ROC-AUC can look strong even when the positive class is extremely rare, because false positives don't move the false positive rate much in absolute terms when there are a lot of negatives. In those settings, PR-AUC is usually more informative because it focuses on precision as recall changes and when positives are rare, that's the trade-off that actually matters operationally.
I still don't treat either as "the" answer in isolation. I pair it with threshold-based metrics and slice analysis because aggregate metrics can hide problems in specific subgroups that matter a lot to the product or to regulators.
Follow-up to expect:
"What's your approach to choosing an operating point once you have the curve?"
8. Class Imbalance: Walk me through your end-to-end approach
What the interviewer is really testing:
Methodical thinking over ad hoc fixes. There are a lot of ways to handle imbalance, and the right approach depends on context.
Sample answer:
First I confirm it's truly an imbalance in the label distribution and not a sampling artifact from how data was collected. Then I pick metrics that won't mislead me like precision/recall style metrics, not accuracy.
Then I decide whether to address it in the data, in the loss function, or at decision time: reweighting classes, over- or under-sampling, or adjusting thresholds post-training. For severe imbalance, I like a staged approach, a high-recall candidate model that casts a wide net, then a precision-focused second stage, plus human review where the stakes justify it. I also make sure my evaluation uses stratified splits so the class distribution is representative in validation.
Follow-up to expect:
"How do you prevent the imbalance handling itself from introducing leakage?"
Part 3: Data (The Stuff That Actually Breaks Models in Production)
Data questions are often underweighted by candidates who've spent most of their prep time on algorithms. In practice, data problems are what break most models in production. Interviewers know this, and they use these questions to find people who've actually shipped things.
9. Missing Data and Corrupted Fields: What's your playbook?
Sample answer:
I start by asking why it's missing, because that determines whether the missingness itself is informative. A field that's frequently empty for one user segment might be a signal, not noise.
For numeric fields, I'll often impute with robust statistics or a model-based imputer. For categoricals, I add an explicit "unknown" bucket rather than trying to guess. For corrupted values, I log the rules, fix the source if possible, and add validation tests so the problem doesn't silently recur. The thing I'm most careful about: imputation that introduces leakage by fitting on the full dataset instead of train-only. That's a surprisingly common mistake that can inflate evaluation metrics in ways that don't survive deployment.
Follow-up to expect:
"When is deletion a better strategy than imputation?"
10. Regularization - L1 vs. L2: What actually changes in the model's behavior?
Sample answer:
L2 tends to shrink weights smoothly toward zero without eliminating them. It's effective when you believe many features contribute a small amount each. L1 can drive weights to exactly zero, which makes it behave like feature selection, useful when you expect sparsity or need a simpler, more interpretable model.
I don't have a default preference. I pick based on a hypothesis about the data's structure and validate it. Worth noting: "regularization" isn't just a hyperparameter knob. Sometimes the best regularization you can do is preventing leakage and cleaning your labels. That's boring but it's often more impactful than tuning lambda.
Follow-up to expect:
"How does regularization interact with collinear features?"
Part 4: Modeling - Trade-offs Over Definitions
These questions test whether you're a practitioner or a textbook reader. The answers below deliberately lead with trade-offs and constraints rather than definitions, because that's what experienced interviewers want to hear.
11. Gradient Descent: Explain it and tell me what you monitor while training
Sample answer:
Gradient descent iteratively updates parameters in the direction that reduces loss. The practical part is knowing what to watch while it runs: loss curves on both train and validation, gradient norms if you're seeing instability, and whether training is compute-bound or input-pipeline-bound.
Learning rate schedules matter a lot, too high and you bounce around without converging; too low and you waste compute on a shallow descent. If training is slow, I profile the input pipeline before I assume the model is the bottleneck. In my experience, it's the pipeline more often than people expect.
Follow-up to expect:
"How do you diagnose whether the input pipeline is the bottleneck?"
12. Linear vs. Logistic Regression: Assumptions, Interpretation, and Failure Modes
Sample answer:
Linear regression predicts a continuous value and assumes a linear relationship in the feature space, after whatever transformations you apply. Logistic regression outputs a probability via a logistic link and is used for classification; coefficients are interpretable in terms of log-odds.
I use logistic regression as a standard baseline for classification problems because it's fast, debuggable, and well-understood. If it underfits, I have clear next steps. The failure modes worth knowing: logistic regression struggles with perfect separation (coefficients diverge), is sensitive to outliers in the feature space without robust preprocessing, and needs calibration if you care about the actual probability values, not just the rank ordering.
Follow-up to expect:
"When does logistic regression give you badly calibrated probabilities?"
13. Decision Trees vs. Random Forests vs. Gradient Boosting: When do you pick each?
Sample answer:
A single decision tree is interpretable and fast to train but notoriously unstable, small changes in data can produce very different trees. Random forests reduce that variance through bagging and tend to behave well out of the box with minimal tuning. Gradient boosting often gives top-tier accuracy on tabular data, but it's more sensitive to hyperparameters and can amplify leakage or drift problems if you're not careful.
My choice is constraint-driven. If I need to explain every decision to a non-technical stakeholder, I'll lean toward a tree or a simple ensemble. If inference latency is tight, gradient boosting might be ruled out. If the data distribution shifts frequently in production, simpler and more stable models are often more defensible. I try to match complexity to the operational context, not just the accuracy benchmark.
Follow-up to expect:
"What's the fundamental difference between bagging and boosting in terms of what errors they fix?"
14. K-Means Clustering: Explain it and how you choose k responsibly
Sample answer:
K-means assigns observations to the nearest centroid and iteratively updates centroids to minimize within-cluster variance. The important caveats: it's sensitive to scale (distance-based, so you need to normalize), it assumes roughly spherical clusters, and results can vary significantly based on initialization.
For choosing k, I'll use elbow-style diagnostics as a starting point, but I don't trust them blindly. What I care more about is whether the resulting clusters actually mean something useful in context, do they produce actionable segments? Are they stable across subsamples? If the data genuinely doesn't fit k-means assumptions, I'll say so and propose alternatives like DBSCAN or hierarchical clustering.
Follow-up to expect:
"How do you handle initialization sensitivity in practice?"
15. PCA and Dimensionality Reduction: When does it help and when does it hurt?
Sample answer:
PCA helps when features are highly correlated and I want a compact representation, faster downstream models, or noise reduction. It hurts when I need interpretability, principal components are linear combinations of the original features and often don't correspond to anything meaningful in the domain. It also creates problems when applied to the full dataset before splitting, which leaks test-set information into the transformation.
I treat PCA as part of a pipeline that's fit on train data only, validated honestly, and documented clearly, especially around the variance explained threshold I chose and why. If someone asks me later why I kept 15 components, I should have a better answer than "that's what I tried."
Follow-up to expect:
"Would you use PCA before a tree-based model? Why or why not?"
16. Feature Engineering Without Leaking the Label
What the interviewer is really testing:
This is a senior filter. Data leakage is one of the most common reasons a model looks great in development and fails in production.
Sample answer:
My rule is simple and I treat it like a contract: every feature must be computable at prediction time using only data that was available at that exact moment, with no information from the future. If I'm using aggregations, they have to be built from historical windows that close before the prediction timestamp. Target encoding gets special treatment, done on folds, not globally. And if the team has had repeated leakage incidents, I'll add automated "could we compute this in production?" checks to the pipeline.
Follow-up to expect:
"Give me an example of a leaky feature you've caught."
17. Feature Importance: How do you explain what's driving predictions?
Sample answer:
I separate global understanding from local explanations. For global importance, I lean toward permutation-based importance because it's model-agnostic and tied directly to performance degradation, which is a more honest measure than impurity-based importance. For local explanations at the instance level, I'll use appropriate methods if they're justified by the use case.
I also sanity-check importance results against leakage suspicion. If a feature looks disproportionately powerful, I verify it's legitimate before shipping. "Too good to be true" is often exactly that.
Follow-up to expect:
"When can feature importance metrics actively mislead you?" (Correlated features, unstable rankings across subsamples.)

18. Data Leakage: Name the five most common types and how you prevent them
What the interviewer is really testing:
Whether you've shipped something and dealt with the consequences. People who've been burned by leakage tend to have very specific answers here.
Sample answer:
The five I watch for most carefully: (1) time leakage - training data includes information from the future; (2) target leakage - features derived from or correlated with the label in ways that wouldn't exist at prediction time; (3) split leakage - the same entity (user, customer, session) appears in both train and validation; (4) preprocessing leakage - scalers, imputers, or encoders fitted on the full dataset before splitting; and (5) proxy leakage - a feature that seems innocuous but encodes the label indirectly, like "refund issued" as a predictor of "fraudulent transaction."
Prevention is mostly process: correct split strategy, pipelines fit on train-only, and a systematic "what would this value be at prediction time?" audit for every feature.
Follow-up to expect:
"How would you detect leakage after the fact if a model is already deployed?"
19. Probability Calibration: When do calibrated probabilities matter?
Sample answer:
If the output probability is driving a downstream decision, pricing, risk thresholds, ranking, loan approval - calibration matters a lot. If you just need a rank ordering and won't be acting on the absolute probability value, a well-discriminating uncalibrated model might be fine.
I validate calibration with reliability diagrams and decision-based backtests: "If we act at 0.8, are outcomes close to 80%?" If it's off, I'll fit a calibration layer on a held-out set and monitor stability over time, because calibration tends to drift along with the model's operating environment.
Follow-up to expect:
"How does calibration change when concept drift occurs?"
Part 5: Production ML (The Questions That Actually Filter Senior Candidates)
This is the section most articles skip or skim. It's also the section that's increasingly prominent in mid-to-senior ML loops, because companies have learned the hard way that a model that can't survive production isn't a model, it's a science project.
20. Model Performance Drops in Production: Debug Drift vs. Bugs
What the interviewer is really testing:
Systematic triage. "Retrain it" is not an answer.
Sample answer:
I start with the boring, high-probability checks: pipeline errors, schema changes, null spikes, and obvious distribution shifts in the features. Then I compare training vs. production distributions to separate a data bug from something more fundamental.
If I suspect drift, I separate data drift, the input distribution changed, from concept drift, where the relationship between features and target has changed. Those have different remedies. Data drift might mean retraining on more recent data. Concept drift might require re-labeling, rethinking features, or questioning the problem framing entirely. Short-term, I'll often roll back or shadow-route to the previous model while I investigate. Longer-term, I set up retraining triggers and tighter monitoring so the next drop gets caught earlier.
Follow-up to expect:
"What monitoring would have caught this sooner?"
21. Monitoring an ML System: What are the three layers?
Sample answer:
I think about monitoring in three layers. First, system health: latency, error rates, and throughput - the infrastructure layer. Second, data health: schema validation, missingness rates, and feature distribution drift - the input layer. Third, model and decision health: prediction distribution, calibration, business KPIs, and ground-truth evaluation when labels eventually arrive.
The part that's often missed: monitoring without runbooks is theater. Every alert should have a corresponding action like rollback, retrain, notify a human, or acknowledge the shift. If you can't specify what you'd do when the alert fires, you haven't finished the monitoring design.
Follow-up to expect:
"What do you track when labels are delayed or expensive to collect?"
22. Feature Stores: What are they and when do teams actually need one?
Sample answer:
A feature store centralizes feature definitions and serving so that training and inference use exactly the same computed values - with versioning, lineage, and reuse baked in. The core value proposition is consistency (fewer train/serve skew bugs), speed (stop rebuilding the same features in six different pipelines), and governance (know what's being used where).
I only advocate for it when the team has reached a scale where feature reuse and consistency pain are genuinely real. If you have two engineers and three models, a feature store is overhead. If you have fifteen engineers managing dozens of models across multiple products, it's almost certainly worth the investment. I've seen teams push for feature stores too early and create complexity they weren't ready to maintain.
Follow-up to expect:
"What are the failure modes of a poorly implemented feature store?"
23. A/B Testing Models: Guardrails, Rollout, and Interpretation
Sample answer:
I define success metrics that reflect both model quality and business impact before the experiment starts, not after. Then I set guardrails: if latency spikes, error rate increases, or any business metric I care about drops by more than a pre-defined threshold, the test stops and I roll back. No exceptions.
For rollout, I usually start with a small percentage of traffic - 5 to 10%, and ramp over time, watching for novelty effects and seasonality. Something that looks like a lift on day one can evaporate by day five. I also decide upfront what I'll do if offline metrics and online metrics disagree, because they often do, and not having a prior answer to that question leads to bad post-hoc reasoning.
Follow-up to expect:
"What if the new model improves clicks but hurts longer-term retention?"
Part 6: System Design - Thinking at Scale
System design questions are where senior candidates earn the title. The goal isn't to produce the perfect architecture, it's to demonstrate that you can think through constraints, trade-offs, and failure modes at the system level.
24. Design a Recommendation System End-to-End
What the interviewer is really testing:
Structured thinking, not encyclopedic knowledge. Use a framework and show you know where the hard problems are.
Sample answer:
I'd start by clarifying the optimization target: engagement, retention, revenue, or something else and identify the online business metric plus the guardrails I won't sacrifice. Then I'd outline the architecture in two stages: candidate generation (fast, approximate retrieval at scale) and ranking (heavier model, personalized scoring). Then re-ranking for diversity and business rules.
Key problems I'd call out explicitly: cold start for new users and new items, exploration vs. exploitation (recommending only what we know the user likes versus introducing variety), and feedback loops that can entrench biases over time. On the production side: feature pipelines, latency budgets at each stage, how the model gets retrained and on what cadence, and what monitoring would catch recommendation quality degradation before users start complaining.
Follow-up to expect:
"How would you prevent the system from creating filter bubbles?" / "How do you evaluate this offline when you can't know what the user would have clicked on if you'd shown something different?"
Part 7: Responsible AI (The Question That's Become a Serious Filter)
Five years ago, this question was rare in technical interviews. Today it's standard at any company that's been through a public incident, a compliance review, or an enterprise sales process. Be prepared to give a substantive answer.
25. Fairness, Bias, and Risk: How do you think about responsible ML?
Sample answer:
I approach it as engineering risk management. The first step is defining what "harm" actually means for this specific product and who could be disproportionately affected. "Bias" is not a single thing, it depends heavily on the context, the user population, and the decisions the model is influencing.
From there, I choose measurements and mitigations that match the actual risk level. I document assumptions, data provenance, and known failure modes so the business can make informed decisions about what to ship and when. If the system affects access to meaningful opportunities, credit, employment, healthcare, I'm stricter about pre-launch review, bias audits, and ongoing monitoring.
And I'd be honest with you: this sometimes creates tension with timelines. My job in those situations is to make the risk visible and concrete, not to make the decision for leadership. But I also won't sign off on something I believe has a meaningful probability of causing harm that we haven't acknowledged or mitigated.
Follow-up to expect:
"What would you do if a model performs well on average but significantly worse for a specific demographic?" / "What if leadership wants to ship despite known issues?"
How to Use This List - A Final Word
Don't memorize these answers. Read them once, understand the structure, and then practice articulating them in your own voice. The version that gets you hired is the one that sounds like you, not like a prep guide.
The pattern you should internalize from every answer above is this: definition → diagnostic process → decision → production consequence. That four-step structure is what separates a strong answer from a textbook answer in every single category on this list.
The candidates I've seen get hired at top ML teams aren't the ones with the most impressive credentials. They're the ones who can walk into an ambiguous problem, name what's uncertain, propose a reasonable path, and explain what they'd watch for to know if it was working. That's what these questions are ultimately testing, and that's the skill worth developing.
Production debugging flow
A quick diagnostic flow engineers follow when model performance suddenly drops in production.
Frequently Asked Questions
What are the most common machine learning interview questions?
The most common machine learning interview questions usually cover model fundamentals, evaluation, data handling, and practical tradeoffs. Candidates are often asked about bias vs variance, overfitting vs underfitting, precision and recall, confusion matrices, cross-validation, regularization, feature engineering, class imbalance, and the difference between common model families like linear models, tree-based models, and ensemble methods. In stronger interviews, especially for mid-level or senior roles, recruiters also ask about production issues like data drift, monitoring, and debugging a model after deployment.
What ML topics should I review before an interview?
Before a machine learning interview, review five core areas. First, fundamentals like supervised vs unsupervised learning, bias, variance, regularization, and gradient descent. Second, evaluation concepts such as precision, recall, F1 score, ROC-AUC, PR-AUC, and calibration. Third, data topics like leakage, missing values, class imbalance, and feature engineering. Fourth, modeling topics including regression, decision trees, random forests, boosting, clustering, and PCA. Fifth, if you are interviewing for ML engineering roles, review production topics such as monitoring, drift, train/serve skew, retraining triggers, and system design.
How do FAANG companies interview machine learning engineers?
Large tech companies typically interview machine learning engineers through a mix of rounds rather than a single pure ML test. You can usually expect some combination of coding, machine learning fundamentals, applied modeling judgment, system design, and behavioral questions. Early rounds often check whether you understand core ML concepts clearly. Later rounds usually push harder on tradeoffs, experimentation, data quality, production thinking, and how you would design or debug a real machine learning system at scale.
What machine learning system design questions are asked in interviews?
Machine learning system design questions usually focus on how you would build, evaluate, deploy, and monitor an end-to-end ML product. Common examples include designing a recommendation system, fraud detection pipeline, ranking system, spam classifier, search relevance model, forecasting workflow, or real-time anomaly detection system. Interviewers are usually testing whether you can think through data pipelines, feature generation, model selection, offline and online evaluation, latency constraints, deployment strategy, monitoring, and retraining.