
How Case Interviews Are Scored: What Partners Look For
Mar 1, 2026
Fundamentals · Scoring, Case Interview, Evaluation Criteria
Road to Offer Team
Road to Offer
We built Road to Offer to make deliberate case practice accessible to every candidate — not just those who can afford $200/hour coaching.
- -Strategy consulting background
- -200+ candidates coached
Published Mar 1, 2026
Summary
The real case interview scoring rubric: what partners evaluate, how each dimension is weighted, and how AI scoring maps to real MBB evaluation criteria.Consulting firms score case interviews across five to seven defined dimensions — including problem structuring, quantitative reasoning, business judgment, communication, and synthesis quality. McKinsey's interviewing page describes the evaluation as assessing how candidates "structure tough, ambiguous challenges, identify important issues, deal with the implications of facts and data, formulate conclusions and recommendations, and articulate thoughts." BCG's careers site adds that candidates are assessed on "numerical skills, business knowledge, and communication style" as separate criteria. Bain's case interview preparation guide focuses on whether candidates "make sensible assumptions, do quick math, and build constructively" — emphasizing practical judgment alongside analytical rigor.
Most candidates have only a vague sense of how they are being evaluated. They know "structure, math, communication" matter, but they don't know the specific dimensions partners use, how those dimensions interact, or which failures are recoverable and which are fatal. This guide breaks down each dimension, explains how McKinsey, BCG, and Bain weight them differently, and maps them to CaseInterviewAI's scoring system so you can practice with the same criteria real partners use.
The 7 Dimensions Partners Actually Evaluate
Every case interview evaluation form covers the same core areas, even though firms use slightly different terminology and weighting. Here is what you are actually being scored on, in order of typical weight.
Dimension 1: Problem Structuring and MECE Quality
Weight: Highest across all firms (approximately 25% of overall evaluation)
This is consistently the most heavily weighted dimension at McKinsey and a top-two dimension at BCG and Bain. It measures whether you can take a complex, ambiguous business problem and decompose it into a clear, logical, mutually exclusive and collectively exhaustive structure.
What strong looks like:
- Buckets are specific to the case, not generic. Instead of "revenues, costs, market," you present "revenue per product line, volume-price mix by channel, competitive pricing pressure in the mid-market segment."
- Every important driver is covered by at least one bucket (collectively exhaustive).
- No overlap or double-counting between buckets (mutually exclusive).
- The structure is organized around an initial hypothesis, not just categorized.
- You can articulate why you chose this structure over alternatives.
What weak looks like:
- Generic frameworks applied without adaptation. If your structure for a healthcare profitability case looks identical to your structure for a retail market entry case, interviewers notice immediately.
- Overlap between buckets. Listing "pricing" and "revenue per customer" as separate branches when one is a component of the other.
- Missing an entire category of relevant drivers.
- No clear starting point or hypothesis to prioritize branches.
At McKinsey specifically: Structure is where McKinsey cases are won or lost. In the candidate-led format, your structure is the roadmap for the entire case. McKinsey's careers page describes the problem-solving interview as evaluating how candidates "structure tough, ambiguous challenges" and "identify important issues" — with structure explicitly listed first. A weak structure derails everything downstream because you don't have an interviewer guiding you through specific questions. McKinsey interviewers sometimes share explicit structure feedback mid-case. If they suggest you "reorganize your approach," your structure has failed.
For a deep dive on building strong case structures, see the profitability framework guide, which covers the most common structuring pattern.
Dimension 2: Quantitative Reasoning
Weight: High (approximately 20% of overall evaluation)
This measures mathematical accuracy, calculation setup, speed under pressure, and your ability to derive business insight from quantitative results. It is not just about getting the right number.
What strong looks like:
- Stating your approach before calculating: "I'll estimate market size as total US households times penetration rate, giving me..."
- Accurate arithmetic under time pressure with appropriate rounding.
- Interpreting results in business context: "That implies a 12% market share, which makes them the number-two player in this market based on what we've discussed."
- Catching your own errors when results don't make intuitive sense.
What weak looks like:
- Math errors that cascade into wrong conclusions.
- Slow calculations that consume disproportionate case time.
- Getting the right number but failing to connect it to what it means for the case.
- Setting up the wrong calculation entirely, which signals a conceptual misunderstanding rather than an arithmetic slip.
Common misconception: Partners don't expect mental math perfection. They expect correct setup and reasonable accuracy. Showing your methodology out loud is more valuable than silent perfect arithmetic, because it demonstrates structured thinking even if you round a number slightly. A candidate who sets up a calculation correctly and says "that gives us roughly $45 million" is evaluated better than one who silently computes $44.7 million but can't explain the approach. BCG's case interview preparation page confirms that BCG assesses "numerical skills" as a distinct dimension — meaning the approach and communication of math is evaluated, not just correctness.
Dimension 3: Business Judgment and Insight Quality
Weight: High (approximately 15% of overall evaluation)
This measures the depth of your analysis: whether you find the real driver, make non-obvious connections between data points, and update your thinking explicitly when evidence warrants it.
What strong looks like:
- Synthesizing multiple data points into a single coherent insight: "Revenue per transaction is flat, but transaction volume is down 18% in the mid-market segment. This suggests customer churn, not pricing, is the primary revenue driver."
- Asking for specific data because of a specific analytical reason, not randomly exploring.
- Demonstrating commercial awareness: understanding industry norms, reasonable margins, competitive dynamics.
- Updating your hypothesis explicitly when data contradicts it: "I initially hypothesized a pricing problem, but this data points to a volume issue. Let me revise my approach."
What weak looks like:
- Collecting data without a clear hypothesis driving the collection.
- Arriving at surface-level insights that any reader of the data could state, without drawing out the business implication.
- Continuing to defend a hypothesis after the data has clearly contradicted it.
- Treating each data point as isolated rather than building a cumulative picture.
For practical techniques on managing hypotheses throughout a case, see hypothesis-driven thinking in case interviews.
Dimension 4: Communication and Delivery
Weight: Moderate-High (approximately 15% of overall evaluation)
This measures clarity of spoken answers, signposting, structured delivery, and active listening. Communication is the dimension that amplifies or undermines every other dimension. A strong analysis delivered poorly scores worse than a slightly weaker analysis delivered with clarity and confidence.
What strong looks like:
- Structuring spoken answers before delivering content: "I see three things in this data. Let me walk through each one."
- Leading with the finding: "The key insight here is X" rather than building to it.
- Using verbal transitions that tell the interviewer where you are in your reasoning.
- Maintaining composure when redirected or challenged, without losing your thread.
What weak looks like:
- Stream-of-consciousness delivery that forces the interviewer to extract insights from a wall of words.
- Going silent for 45+ seconds without narrating your thought process.
- Burying conclusions at the end of long explanations.
- Passive hedging language that undermines confidence: "I guess maybe we could possibly consider..."
For a detailed guide on case interview communication technique, see communication tips for case interviews.
Dimension 5: Creativity and Novel Thinking
Weight: Moderate (approximately 10% of overall evaluation)
This dimension has gained weight in recent years as firms increasingly test candidates' ability to go beyond standard framework application. It measures whether you bring original thinking to the problem.
What strong looks like:
- Proposing solutions or angles the interviewer hasn't heard from the last 20 candidates.
- Making connections across industries or domains: "This pricing dynamic is similar to what airlines did with dynamic pricing. Could the client use a similar tiered approach?"
- Challenging assumptions in the case prompt when appropriate.
- Offering a genuinely differentiated recommendation rather than the textbook answer.
What weak looks like:
- Every answer follows the same template regardless of context.
- Inability to think beyond the framework you laid out at the beginning.
- Answers that are technically correct but boring: they solve the math without adding strategic value.
Dimension 6: Hypothesis Management and Process
Weight: Moderate (approximately 10% of overall evaluation)
This measures whether you approach the case with a clear hypothesis, test it systematically, and update it based on evidence. It is closely related to business judgment but focused specifically on your analytical process rather than the quality of individual insights.
What strong looks like:
- Forming a specific initial hypothesis before diving into analysis.
- Explicitly stating what data would confirm or disconfirm your hypothesis.
- Pivoting cleanly when evidence points in a different direction.
- Maintaining a clear analytical thread throughout the case so the interviewer can follow your logic.
What weak looks like:
- "Boiling the ocean" — exploring data randomly without a hypothesis guiding your choices.
- Asking for data without articulating why you need it.
- Failing to update your hypothesis even when contradictory evidence is presented.
- Losing track of where you are in your own framework.
Dimension 7: Synthesis and Recommendation Quality
Weight: Moderate (approximately 5% of overall evaluation, but high impact on final impression)
The synthesis is your last impression. It measures whether your final recommendation is clear, actionable, well-supported by the analysis you've done, and delivered with appropriate confidence. Despite its relatively lower weight in formal scoring, a strong synthesis can elevate a borderline case performance, and a weak one can undermine a strong performance.
What strong looks like:
- Bottom-line-first delivery: "My recommendation is X, for three reasons."
- Specific and actionable: "Cut operating costs in the European manufacturing network by consolidating three plants into one, targeting $80-100M in annual savings" rather than "reduce costs."
- Acknowledging key risks and how you would manage them.
- Connecting the recommendation back to the case objective stated at the beginning.
What weak looks like:
- Hedged or indecisive conclusions: "I think maybe they should consider possibly entering the market..."
- Recommendations that don't match the data presented during the case.
- Missing the "so what" — stating findings without a clear action implication.
- Rushing through the synthesis because you ran out of time.
For a complete guide on delivering strong case recommendations, see case interview synthesis.
How McKinsey, BCG, and Bain Weight Dimensions Differently
The evaluation dimensions are consistent across firms, but each firm emphasizes different areas based on their interview format and culture.
McKinsey weights structure and synthesis highest. The candidate-led format demands that you produce a usable analytical roadmap at the start and a clear, confident conclusion at the end. McKinsey's interviewing page lists the ability to "structure tough, ambiguous challenges" as the first evaluation criterion, confirming that structure is the primary differentiator. McKinsey also evaluates hypothesis management more explicitly than other firms because in a candidate-led case, your hypothesis determines where you direct the entire analysis. A weak hypothesis means wasted case time.
BCG weights insight quality and quantitative reasoning particularly highly. BCG cases are exhibit-heavy. BCG's interview preparation guide explicitly names "numerical skills, business knowledge, and communication style" as the three scored criteria for client-facing roles. The ability to extract a non-obvious insight from a visualization and connect it to a quantitative business implication is what BCG is looking for.
Bain weights communication and composure relatively higher than the other two. Bain's case interview page emphasizes that candidates should "make sensible assumptions, do quick math, and build constructively on others' ideas" — framing the evaluation around practical collaboration rather than formal analytical precision. Bain interviewers report caring more about whether you would be effective in a real client conversation than whether your structure is formally MECE in every detail.
For firm-specific guidance, see the McKinsey case interview guide and common case interview mistakes.
How Our AI Scoring Maps to the Real Rubric
CaseInterviewAI's debrief system is designed to reflect the same dimensions MBB interviewers evaluate. Here is the mapping between our AI scoring categories and what partners assess:
| Our Scoring Dimension | Weight | Maps to Partner Rubric |
|---|---|---|
| Structure & MECE | 25% | Problem Structuring |
| Math & Quantitative | 20% | Quantitative Reasoning |
| Business Judgment | 15% | Insight Quality |
| Communication & Signposting | 15% | Communication / Delivery |
| Creativity | 10% | Novel Thinking |
| Hypothesis-Driven | 10% | Hypothesis Management |
| Synthesis | 5% | Recommendation Quality |
These weights are calibrated based on coaching experience and the dimensions that most consistently differentiate candidates who receive offers from those who don't. They are not official MBB weights, which are not publicly published.
Want to see how you score across all 7 categories? Try a free case interview to get your baseline scores and identify which dimensions to prioritize in your preparation.
The Scoring Scale: What the Numbers Mean
Most MBB firms use a 4-point or 5-point scale applied to each dimension:
| Score | Label | What It Means |
|---|---|---|
| 1 | Insufficient | Fundamental lack of the skill. Almost always disqualifying. |
| 2 | Adequate | Below the hiring bar. Shows the basics but not at the level required. |
| 3 | Good | Meets the hiring bar. Solid performance with minor areas for improvement. |
| 4 | Very Good / Excellent | Above the bar. Strong, differentiated performance. |
Scores of 1 or 2 on any key dimension typically lead to rejection, even if other dimensions are strong. The threshold for an offer is generally a 3 or above across all dimensions, with at least one or two dimensions at a 4 to demonstrate standout potential.
In final rounds with senior partners, the evaluation often shifts toward a simpler binary: hire or no-hire. But partners still reference the dimensional criteria when discussing candidates in calibration sessions where multiple interviewers compare their assessments.
What Most Prep Guides Get Wrong About Scoring
Treating all dimensions as equal. They are not. A single arithmetic error mid-case is recoverable. A structuring failure that produces an incoherent roadmap typically prevents you from reaching a good synthesis, which means one failure cascades across multiple dimensions. Know which dimensions are load-bearing.
Ignoring composure. Most guides cover structure, math, and communication but barely mention how you handle pushback. Yet partners deliberately stress-test candidates by challenging their conclusions, presenting contradictory data, or asking "are you sure?" Candidates who have never practiced maintaining their position under pressure are unprepared for one of the most predictable moments in every MBB interview.
Treating the recommendation as an afterthought. The synthesis is where you crystallize everything you've done in the past 25-30 minutes into a clear action for the client. Many candidates run out of time, rush the synthesis, and deliver a vague conclusion, undermining an otherwise solid case performance. Budget 2-3 minutes for synthesis. It should not take 30 seconds.
Over-rotating on math. Quantitative reasoning is important, but it is one of seven dimensions. A candidate with perfect mental math but poor structure will not get an offer. Allocate prep time proportionally across dimensions, not based on what feels most "practiceable."
Not tracking improvement across dimensions. Doing case after case without measuring which specific dimensions are improving and which are stagnating is the most common plateau trap. Management Consulted's case scoring guide notes that McKinsey scores each piece of the case separately — meaning you must "pass" each question, not just the aggregate — while BCG and Bain give more aggregated end-of-case scores. After every practice case, score yourself on each dimension or use a system that scores you automatically. Then target your weakest dimension with specific drills before running more full simulations.
Key Takeaways
- Case interviews are scored across 7 core dimensions: Problem Structuring, Quantitative Reasoning, Business Judgment, Communication, Creativity, Hypothesis Management, and Synthesis.
- Structure failures are the most common reason strong analytical candidates don't get offers. The structure is the case's foundation, and a failure there cascades into every other dimension.
- McKinsey weights structure and synthesis highest. BCG weights insight quality and quantitative reasoning highest. Bain weights communication and composure relatively higher.
- The scoring scale is typically 1-4, where 1 and 2 are below the bar. You need a 3 or above across all dimensions with standout performance in at least one.
- Track your scores across dimensions over time to identify and target your weakest areas rather than doing unfocused full-case practice.
Practice with AI
Sources and Further Reading (checked March 1, 2026)
- McKinsey case interview preparation, problem-solving evaluation criteria: mckinsey.com/careers/interviewing
- BCG case interview preparation, scored dimensions including numerical skills and communication style: careers.bcg.com/global/en/case-interview-preparation
- Bain case interview preparation, evaluation of practical judgment and collaboration: bain.com/careers/hiring-process/case-interview
- Management Consulted, insider look at case interview scoring system including McKinsey question-by-question scoring: managementconsulted.com/insiders-look-case-scoring-system
- IGotAnOffer, BCG case interview guide including firm-specific evaluation emphasis: igotanoffer.com/blogs/mckinsey-case-interview-blog/bcg-case-interview
Frequently asked questions
Continue your prep path
Next actions based on this article: one pillar hub, two related guides, and one conversion step.
Pillar hub
MBB and Firm-Specific Hub
Related guide
Case Interview Communication: How to Think Out Loud
Related guide
Reading Charts and Exhibits in Case Interviews
Related articles
Case Interview Communication: How to Think Out Loud
Learn how to verbalize your thinking in case interviews. Covers signposting, structuring spoken answers, managing silence, and the communication habits top candidates use.
Reading Charts and Exhibits in Case Interviews
How to read charts, graphs, and data tables in case interviews. Covers exhibit types, common traps, a 4-step reading method, and worked examples under time pressure.
Case Interview Synthesis: 60-Second Recommendation Template + Examples
Learn how to deliver a clear recommendation at the end of a case interview using a 60-second synthesis template, with worked examples and common mistakes.