Most AI interview questions test the wrong thing. "What AI tools do you use?" tests adoption. "What is a hallucination?" tests vocabulary. "How do you use AI in your workflow?" invites a rehearsed narrative that tells you what the candidate wants you to hear, not what they actually do.
None of these questions predict whether a candidate will verify AI output before sending it to a client, recognize when data should not enter an AI tool, or exercise judgment about when to override an AI recommendation. And those are the behaviors that determine whether a hire succeeds or fails in an AI-augmented role.
The questions below are designed to reveal AI judgment: the habitual decision-making patterns that govern how a person actually works with AI under realistic conditions. They are organized by the five dimensions of AI readiness, with scoring guidance for each.
Before you start: why the question format matters
The difference between a knowledge question and a judgment question is structural.
A knowledge question has a correct answer: "What is a hallucination?" The candidate either knows the definition or does not. This tells you about their vocabulary, not their behavior. A candidate who can define "hallucination" may still accept hallucinated content at face value in their work, because 38% of business executives have done exactly that (Deloitte, 2024).
A judgment question presents a situation with competing considerations (time pressure, authority, stakes, ambiguity) and asks the candidate to describe their response. There is no single correct answer. What you are listening for is the pattern of the response: what does the candidate notice? What do they prioritize? What risks do they identify? What do they miss?
As we explored in our analysis of why "I use ChatGPT" tells you almost nothing about a candidate, self-reported AI proficiency is systematically unreliable. These questions are designed to move past self-report into observable judgment patterns.
Dimension 1: AI fluency
Question 1: "You need to prepare a competitor analysis for a client meeting tomorrow. You have three hours. Walk me through your approach, including how and where you would use AI."
What to listen for: Does the candidate describe a process that includes both generation and verification? Do they mention checking AI-generated claims against primary sources? Do they distinguish between what AI can do well (structure, synthesis, first drafts) and what requires human judgment (strategic interpretation, client-specific context, accuracy verification)?
Strong signal: The candidate describes using AI for the initial draft or research synthesis, then spending meaningful time verifying specific claims, adding proprietary knowledge, and tailoring the analysis to the client's context. They mention checking at least one factual claim against an original source.
Weak signal: The candidate describes generating the analysis with AI and presenting it, with no mention of verification. Or they describe doing the entire analysis manually, suggesting they are not leveraging AI productively. Both extremes, uncritical dependence and complete avoidance, are informative.
Question 2: "Describe a time when AI saved you significant time on a work task. What was the task, and what did you do differently because AI was involved?"
What to listen for: Specificity. A strong answer describes a concrete task, what the AI produced, and what the candidate did with the output. A weak answer stays general: "I use AI all the time to be more productive." The follow-up question, "what did you do differently?", reveals whether the candidate adjusted their process (for example, spending saved time on verification or strategic thinking) or simply produced more output faster without changing their approach to quality.
See the gap for yourself
Take the free Aptivum Snapshot (10 questions, 8 minutes) and find out where you actually stand on AI readiness.
Dimension 2: Critical evaluation
Question 3: "An AI tool generates a report for you that includes three statistics you have not seen before. They look plausible. What do you do before using them in a client deliverable?"
What to listen for: This is the single most diagnostic question in the set. The candidate's instinct, verify or proceed, reveals their critical evaluation habit more reliably than any other single question. Do they describe checking the statistics against primary sources? Do they mention that AI can fabricate citations and statistics? Do they have a specific verification method?
Strong signal: The candidate immediately describes a verification step: checking the original source, searching for the statistic independently, or cross-referencing with known data. They may mention that AI-generated statistics should be treated as unverified by default.
Weak signal: The candidate says they would "review" the statistics but cannot describe how. Or they say they would trust the statistics because the AI tool is generally reliable. The absence of a concrete verification method is the red flag, as it suggests the candidate does not have a verification habit.
Question 4: "Have you ever caught an error in AI-generated content? Tell me about it."
What to listen for: A specific example. Candidates with genuine critical evaluation habits will have stories: a hallucinated citation, a factual error, an outdated statistic, a recommendation that did not fit the context. Candidates without this habit will struggle to provide an example, because they have not been looking for errors. The inability to recall a single instance of catching an AI error is itself a data point: it suggests the candidate has not been verifying.
Dimension 3: Ethics and data privacy
Question 5: "You want to use AI to draft personalized outreach emails to 50 candidates. Your CRM contains their names, salary expectations, and interview notes from previous rounds. How do you proceed?"
What to listen for: Does the candidate recognize the data sensitivity issue before being prompted? Do they distinguish between information that can safely enter an AI tool and information that cannot? Do they consider whether the AI platform is company-approved versus public? 57% of enterprise employees have entered confidential information into public AI tools (TELUS Digital, 2025). This question reveals whether the candidate would be in that 57% or not.
Strong signal: The candidate identifies the sensitivity of salary expectations and interview notes without prompting. They propose anonymizing or abstracting the data, using a company-approved AI platform, or structuring the task so that sensitive details remain out of the AI tool.
Weak signal: The candidate describes pasting the CRM data into an AI tool to generate the emails, with no mention of data sensitivity. This does not mean they are a bad candidate; it means they have a specific, addressable gap in data privacy awareness.
Question 6: "A colleague shares an AI-generated analysis that includes a recommendation you believe is based on biased training data. How do you handle it?"
What to listen for: This question tests both ethical awareness and interpersonal judgment. Does the candidate recognize the bias concern as legitimate? Do they raise it with the colleague? Do they escalate? Do they propose a way to check whether the bias affected the outcome? The social dimension matters, as many candidates will know that AI bias exists in theory but will hesitate to raise the concern with a colleague in practice.
Dimension 4: Judgment
Question 7: "Your manager asks you to use AI to analyze customer feedback data that includes personal identifiers and complaints about a safety issue. The deadline is this afternoon. What do you do?"
What to listen for: This is a conflict-of-priorities question. Speed (the deadline), authority (the manager), data sensitivity (personal identifiers), and seriousness (safety issue) all compete. The candidate's response reveals their judgment hierarchy. Do they prioritize the deadline or the data concern? Do they push back on the manager, or comply? Do they recognize that a safety-related analysis has higher stakes than a routine report?
Strong signal: The candidate addresses the data sensitivity first, proposing anonymization or use of an approved tool, and then addresses the deadline. They may push back on the timeline or propose a modified approach. They recognize that the safety dimension raises the stakes.
Weak signal: The candidate focuses on meeting the deadline and describes completing the task as requested, without raising data or safety considerations. Compliance without consideration is the pattern that predicts poor judgment in high-stakes situations.
Question 8: "Give me an example of a situation where you decided NOT to use AI for a task. Why?"
What to listen for: This is perhaps the most underrated question in AI judgment assessment. The ability to articulate when AI is inappropriate is a stronger signal of judgment than any description of AI use. Does the candidate have a framework for deciding when AI adds value and when it introduces risk? Or do they use AI for everything, or for nothing?
Strong signal: The candidate describes a specific situation (sensitive data, high stakes, need for original thinking, regulatory context) and explains why AI was not the right tool. The reasoning reveals their judgment framework.
Weak signal: The candidate cannot think of an example, suggesting they have not developed the habit of evaluating whether AI is appropriate for a given task. Alternatively, they describe avoiding AI entirely, which suggests avoidance rather than judgment.
Dimension 5: Human-AI collaboration
Question 9: "You and a team member are jointly producing a report. They hand you their section, which they tell you was drafted with AI assistance. You notice several claims that feel unsupported. What do you do?"
What to listen for: The interpersonal dimension. Does the candidate verify the claims independently? Do they raise the concern with the colleague? How do they navigate the social dynamics of questioning AI-assisted work without questioning the colleague's competence? This question reveals whether the candidate can maintain professional relationships while upholding quality standards in AI-augmented collaboration.
Question 10: "How do you think about the division of work between you and AI tools? What stays with you, and what do you delegate?"
What to listen for: A thoughtful framework. Strong candidates have an articulated, even if informal, model for what AI does well and what requires human judgment. They can describe tasks they delegate to AI (first drafts, synthesis, formatting) and tasks they keep (strategic interpretation, ethical judgment, client-specific context, final verification). The specificity of the framework matters more than the framework itself.
Two meta-questions for senior roles
Question 11: "If you were onboarding a new team member, what would you tell them about using AI responsibly in this role?"
What to listen for: This question tests whether the candidate can articulate AI judgment principles to others, which is a higher-order skill than applying them personally. Strong candidates describe specific guidance: verify output, protect sensitive data, calibrate AI use to stakes, maintain human oversight. This question is particularly valuable for management and team lead roles, where the candidate will influence how others use AI.
Question 12: "How do you stay current on the limitations and risks of AI tools you use?"
What to listen for: Ongoing learning habits. AI capabilities and limitations change rapidly. A candidate who treats their AI knowledge as fixed ("I took a course last year") is different from one who actively tracks changes, reads about failures, and adjusts their behavior as tools evolve. With the EU AI Act's Article 4 requiring ongoing AI literacy and research showing that AI competence decays within three to four months without reinforcement, this is not a nice-to-have. It is a regulatory and practical necessity.
Scoring guidance
When evaluating responses, resist the temptation to score based on whether the candidate gives the "right" answer. There is no right answer to most of these questions. Instead, score on three observable dimensions:
Awareness: Does the candidate identify the key considerations in the scenario without being prompted? A candidate who spontaneously raises data privacy concerns in question 5 demonstrates a different level of awareness than one who addresses it only when asked directly. Unprompted identification of risks is a stronger signal than prompted acknowledgment.
Specificity: Does the candidate describe concrete actions, or do they speak in generalities? "I would verify the statistics" is a general answer. "I would search for the original source, check whether the citation exists, and cross-reference the number against industry benchmarks I already know" is a specific one. Specificity correlates with habitual practice: candidates who actually verify AI output can describe their process. Candidates who do not have a verification habit speak in abstractions.
Trade-off reasoning: Does the candidate acknowledge competing considerations? The best AI judgment is not about always choosing caution over speed, or privacy over productivity. It is about recognizing that these trade-offs exist and making a deliberate choice. A candidate who says "I would miss the deadline to protect the data" demonstrates different judgment than one who says "I would anonymize the data so I can meet the deadline and protect privacy." Both may be appropriate depending on context. The candidate who does not recognize that a trade-off exists is the one to be concerned about.
What these questions do not tell you
Interview questions, even good ones, have structural limitations. They test verbal reasoning about hypothetical situations. They do not test behavior under actual working conditions. For a deeper look at how scenario-based assessment complements interview questions, see how scenario-based assessment reveals real AI judgment. A candidate who gives a thoughtful answer about data privacy in an interview may still paste confidential data into a public AI tool when facing a real deadline, because habitual behavior under pressure does not always match deliberative reasoning in a low-stakes setting.
This is why interview questions and behavioral assessment are complements, not substitutes. Interview questions reveal the candidate's mental model: what they think about, what they prioritize, what risks they recognize. Behavioral assessment reveals what they actually do when placed in a realistic scenario. The combination of both produces a signal that neither can generate alone.
How to use these questions
These questions are not designed to be used as a full battery in a single interview. Select two to four based on the role's risk profile:
For client-facing roles where AI output may reach external audiences, prioritize questions 3, 5, and 7 (critical evaluation, data privacy, and judgment under pressure).
For team lead and management roles, prioritize questions 9, 11, and 8 (collaboration, mentoring, and knowing when not to use AI).
For analytical and research roles, prioritize questions 1, 3, and 10 (productive AI use, verification habits, and work division framework).
For any role involving sensitive data (HR, finance, legal, healthcare), questions 5 and 7 are non-negotiable.
The questions work best when followed up with probing: "Tell me more about how you would verify that." "What would you do if your manager pushed back?" "How would you explain that to a client?" The follow-up reveals depth. The initial answer reveals instinct.
For a deeper exploration of behavioral patterns that indicate poor AI judgment, see red flags in AI readiness: 5 patterns that predict poor AI judgment.
Want to go beyond interview questions? The Aptivum assessment measures all five dimensions through scenario-based evaluation, giving you a scored profile, not just an interview impression. Take the free Snapshot to see it in action.


