Google's famous interview questions have been all over the internet, with many of these weird Google interview questions becoming legendary in the tech industry. We found a list compiled by Lewis Lin, a Seattle based job coach. It was fun to read through their questions; we've pulled out some of our favorites here:
- If you look at a clock and the time is 3:15, what is the angle between the hour and the minute hands?
- Design an evacuation plan for San Francisco.
- How many vacuum’s are made per year in USA?
- How many piano tuners are there in the entire world?
- How many haircuts do you think happen in America every year?
- How much should you charge to wash all the windows in Seattle?
-
Explain the significance of "dead beef"
Google's analysis – How they did it
Google took data from tens of thousands of interviews and compared how candidates in those interviews score relative to how the hired candidates actually performed in their job. They looked for correlations between scores on specific questions and the overall score provided by specific interviewers to see whether there were people at Google who were particularly good at hiring.
Google's analysis – What they found
Google took data from tens of thousands of interviews and compared how candidates in those interviews score relative to how the hired candidates actually performed in their job.As mentioned before, Google found zero correlation between how well a candidates scored on brain teaser questions and how well they performed in their job.
They also found that only one interviewer was consistently good at at predicting whether a candidate would perform well. Think about that for a moment. Across thousands of interviews, Google learned that all but one of their interviewers were not predictive—you could essentially flip a coin
What we can learn
Across thousands of interviews, Google learned that all but one of their interviewers were not predictive—you could essentially flip a coinBeyond Brain Teasers: Google's Shift to Behavioral Interviews
After discovering their traditional approach was ineffective, Google pivoted toward more structured interviewing methods. Today, Google behavioral interview questions focus on past experiences and specific situations rather than abstract puzzles. Google case study interview questions now emphasize real-world problem-solving scenarios relevant to the actual role. This shift demonstrates how even the most innovative companies must evolve their hiring practices based on data-driven insights about what actually predicts job success.
One of the biggest takeaways from Google's story is that you can actually know, with some sort of certainty, which interview questions are working and which are not. Many hiring managers continue to use the same interview questions over and over—or pick new ones based on recommendations, a recent article read, or how they're feeling on a given day. Google has shown that we can analyze the questions we ask and learn how predictive our questions are.
There are lots of other, specific lessons that can be gleaned from Google's experience, but we'll call out three:
- A bad process is costly: Think about all the costs associated with investing in a process that is not predictive. Time wasted by candidates and interviewers, a bad candidate experience (I guess few people enjoyed these fear-inducing questions), and likely many false positives from basing decisions on a non-predictive indicator
- Causal hypothesis must be validated: Google is a tech giant with a complex product solving big problems. I'm sure Google wants those it hires to be "Smart." What better way to test whether someone is smart than by giving candidates tricky, complicated questions to see how they think? Makes perfect sense. Except the data showed that it didn't. As humans, we're really good at developing plausible-sounding stories. We do this all the time, and we especially do this in hiring. For this reason, it's critical that we run the necessary analysis to test our hypotheses.
- Collecting data on your hiring process is critical: This point is somewhat obvious, but is so critical—and oft overlooked—that I'm going to say the obvious: Google never would never have learned they were wrong if they hadn't collected and stored their interview score data. They would still be wasting time, inducing fear, and hiring false positives.
What assumptions about what accurately predicts a good hire are built into your hiring process? How might you test these? What data do you need to start collecting to run the required analyses? Stay tuned for a post on Journeyfront's tips for running your own analysis on the effectiveness of your hiring process.
STOP GUESSING, START MEASURING
See Which Interview Questions Actually Predict Performance
Journeyfront connects interview scores to post-hire outcomes — so you can see what's working and stop using questions that don't predict success.
Get the Interview Scorecard Guide

