Employers love tricky job interview questions, but they’re actually useless
If you have ever interviewed for a job, there is a non-trivial probability that you have encountered “tricky” or quirky interview questions. These are questions that are intentionally unexpected, abstract, or only loosely related to the actual requirements of the role. Rather than systematically assessing job-relevant skills, they are designed to surprise candidates, test composure, or signal creativity.
Interviewers often defend these questions as clever ways to evaluate problem-solving ability, cultural fit, or performance under pressure. The evidence tells a different story. Decades of research in industrial-organizational psychology show that unstructured, brainteaser-style interviews have low predictive validity. They generate noise, not insight. At best, they measure how comfortable someone is with improvisation. At worst, they measure how similar the candidate is to the interviewer.
Cases in point
To illustrate the point, here are some common examples, ordered from least absurd, or at least somewhat defensible, to most absurd:
1. What is your biggest weakness?
Nominally job-related, though usually answered strategically rather than honestly. The only rational way to respond is to disguise a strength as a flaw. It is less a test of self-awareness than an audition for plausible humility.
2. Sell me this pen.
Some relevance for sales roles, but still an artificial performance detached from real context. Popularized by The Wolf of Wall Street, it reinforces the myth that great sales is about fast talk rather than listening, diagnosing needs, and building trust.
3. Tell me about a time you failed.
In principle, a legitimate behavioral question. In practice, often an invitation to narrate a carefully curated setback that highlights resilience, grit, and eventual triumph. It rewards storytelling ability more than learning agility.
4. How many tennis balls can fit inside a Boeing 747?
A classic “guesstimate” puzzle meant to test structured thinking. Geeks may love it, but it predicts little beyond prior exposure to similar puzzles. If you want to measure cognitive ability, there are far more reliable and validated tools.
5. How many windows are there in New York City?
Same logic, further removed from any realistic job task. For what it’s worth, large language models estimate the number in the tens of millions, depending on assumptions. Which illustrates the deeper point: if ChatGPT can answer it in seconds, why are we using it to judge human potential?
6. If you were an animal, which one would you be and why?
A thinly veiled personality quiz. It feels like a BuzzFeed throwback disguised as talent assessment. The answer often reveals more about the interviewer’s projections than the candidate’s traits.
7. If you could have dinner with any historical figure, who would it be?
A pleasant icebreaker masquerading as a values assessment. It doubles as a signaling exercise: How curious, cultured, contrarian, or provocative can you appear in under 30 seconds? Say Nelson Mandela and you signal virtue. Say Steve Jobs and you signal ambition. Say Machiavelli and you signal strategic depth. But say Stalin and suddenly the interview turns into a moral inquiry. Was that intellectual curiosity, dark humor, or deeply questionable judgment? The question reveals less about your leadership potential than about your risk appetite for reputational self-sabotage.
8. If you were a kitchen utensil, which one would you be?
At this point, the exercise has drifted into sheer parody—shows like The Office come to mind. Spoon suggests reliability. Knife signals edge. Spork implies versatility. The real variable being tested may simply be how badly you want the job, signaled by the fact that you haven’t just walked out of the room.
The science
So, what does the actual science of interviewing say?
First, there is evidence that some interviewers are not merely misguided, but derive a certain Machiavellian pleasure from putting candidates on the spot. Research on interviewer behavior shows that individuals higher in everyday sadism or dominance are more likely to ask stress-inducing or intentionally uncomfortable questions. In other words, the brainteaser may sometimes be less about assessing you and more about interviewers enjoying the deviant power dynamic.
Second, the predictive validity of unstructured interviews is consistently low. Meta-analyses spanning decades show that traditional, free-flowing interviews correlate only modestly with later job performance. The problem is not conversation per se, but inconsistency. Different candidates get different questions. Interviewers rely on intuition. Evaluation criteria shift midstream. The result is noise, bias, and overconfidence, and unfortunately, these issues often go undetected because of the subsequent confirmation bias or failure to admit mistakes by hiring managers. In essence, if an interviewer likes you, they will either continue to like you after you are hired or pretend you are doing a great job to avoid looking like a fool.
By contrast, structured interviews work. The formula is hardly mysterious: define the competencies that matter for the job; ask all candidates the same job-relevant questions; anchor evaluations to predefined scoring rubrics; and combine interview data with other validated predictors such as cognitive ability or work samples. Behavioral questions about past actions and situational questions tied to realistic job scenarios consistently outperform seemingly clever riddles and quirky brain teasers.
The role of AI
And then there is AI, not so much the elephant in the room as the bull in the china shop, already rearranging the furniture while we are still debating the seating plan.
In a world where candidates can rehearse flawless answers with generative tools, the theatrical interview becomes even more obsolete. Chatbots can generate polished responses to “biggest weakness” or “sell me this pen” in seconds. Ironically, the more predictable and formulaic the question, the easier it is to game. This raises the bar for employers: Assessment must shift toward observable skills, simulations, job trials, and multi-source data.
This does not mean interviews become irrelevant. It means they must evolve. When information is abundant and answers are cheap, the premium shifts from rehearsed narratives to demonstrated capability. Instead of asking candidates what they would do, employers can observe what they actually do: solve a real problem, analyze a live case, critique a flawed strategy, or collaborate with a future teammate. AI can help candidates prepare, but it cannot fully fake sustained performance in a realistic simulation.
There is also a deeper irony. The very tools that allow candidates to polish their answers can help employers design better assessments. AI can assist in standardizing questions, generating competency-based scenarios, flagging bias in evaluation, and even predicting which interview questions correlate with outcomes. In other words, AI exposes the weakness of theatrical interviewing while simultaneously offering the tools to fix it. The real risk is not that candidates use AI. It is that employers fail to upgrade their methods accordingly.
In sum, the future of interviewing is not about trickier questions. It is about better design. The uncomfortable truth is that quirky interview questions persist because they are fun, easy, and ego-affirming. But hiring is too important to be left to entertainment. If organizations are serious about talent, they must replace improvisational theatre with evidence-based assessments, and have the humility and self-critical honesty to truly test the outcome of their decisions to acknowledge when they are wrong, and make an effort to tweak things and improve.