1 of 23

Getting AI to Think Critically (Not Just Agree With You)

AI Critical Analysis Reference Guide

2 of 23

Core Concept Overview

What Is AI Critical Analysis?

Using specific prompt techniques to override AI's natural agreeable bias and force it to challenge assumptions, argue contrary positions, reveal blindspots, and predict failure scenarios rather than providing supportive validation.

Why It Works for Business Decisions

Most people accidentally turn AI into an expensive echo chamber by asking leading questions that prime agreeable responses. This leads to validated thinking that feels good but doesn't improve decision quality or reveal critical flaws.

Mental Shift Required

From:

"What do you think about my idea?" (seeking validation)

To:

"What's wrong with my thinking and how could this fail?" (seeking analysis)

3 of 23

3-Step

Critical Analysis Audit

Step 1: Assumption Challenge

Strip away what you "know" to be true about your decision

Identify which beliefs are actually untested assumptions

Ask: "What am I taking for granted that might be wrong?"

4 of 23

Step 2: Contrarian Position

Force AI to argue against your preferred approach

Demand evidence and reasoning, not gentle suggestions

Most decisions fail because teams optimise for feel-good consensus

5 of 23

Step 3: Premortem Validation

Imagine your decision has failed completely

Work backward to identify realistic failure pathways

Focus on what you can prevent now versus what you'll regret later

6 of 23

AI Prompt Library

Prompt 1: Assumption Challenger

Purpose: Break down any business decision to identify and test hidden assumptions.

I'm about to make this decision: [Your specific decision and reasoning]

Your role: Challenge every assumption I'm making. For each assumption you identify:

  1. Label it clearly as an assumption
  1. Rate how confident you are that it's actually true (1-10)
  1. Suggest the cheapest way to test or validate it
  1. Explain what happens to my decision if this assumption is wrong

Be ruthless. I'd rather discover flawed thinking now than fail later.

7 of 23

Sample Input:

50-person SaaS company deciding to spend £200K on customer success automation instead of hiring 3 more sales people, assuming current customers will renew at 85% rate and new automation will reduce churn by 15%.

8 of 23

Prompt 2: Contrarian Analyst

Purpose: Create strong arguments against your position when everyone seems to agree.

I believe [state your position/decision strongly]. You disagree completely.

Make the strongest possible case against my position. Use specific evidence, logical reasoning, and real-world examples. Don't hedge or soften your critique.

Your goal: Convince me I'm wrong. Be as persuasive and thorough as possible in your disagreement.

9 of 23

Sample Input:

"I believe we should launch our new product feature in Q2 because our beta users love it, the market timing is perfect, and our competitors don't have anything similar."

10 of 23

Prompt 3: Blindspot Detector

Purpose: Reveal missing perspectives and unconsidered factors in complex decisions.

Based on this situation: [describe your business context and decision]

What am I probably not seeing? What blindspots do people in my position typically have? What questions am I not asking that I should be?

Focus on what's missing from my analysis, not what's already there. Be specific about:

  • Stakeholders I might be overlooking
  • Market dynamics I'm not considering
  • Internal constraints I'm ignoring
  • Long-term consequences I'm missing

What would someone completely outside my situation notice that I don't?

11 of 23

Sample Input:

Mid-stage startup CEO deciding whether to expand internationally, focusing on market size and product-market fit but may be missing operational complexity, legal requirements, and resource drain on core business.

12 of 23

Prompt 4: Premortem Facilitator

Purpose: Predict realistic failure scenarios before making major commitments.

It's one year from now. My decision to [your decision] has failed spectacularly.

Tell the story of how it failed. Be specific about:

  • What went wrong first
  • How small problems became big problems
  • What warning signs I ignored
  • What I should have done differently
  • Who was affected and how

Make this failure story as realistic and detailed as possible. Don't hold back on the consequences.

13 of 23

Sample Input:

Decided to switch from current project management tool to new platform for 200-person company, expecting 2-week transition and improved productivity, but considering failure scenarios around adoption, data migration, and workflow disruption.

14 of 23

Quick Reference Decision Tree

When facing any business decision where you want critical analysis:

01

Start Here: Am I seeking validation or analysis?

  • If seeking validation → Stop. Rethink your question
  • If seeking analysis → Continue to step 2

02

Do I have strong assumptions about what will work?

  • If yes → Use Assumption Challenger prompt
  • If no → Continue to step 3

03

Does my team/I feel confident this is the right approach?

  • If yes → Use Contrarian Analyst prompt
  • If uncertain → Continue to step 4

04

Am I considering all relevant perspectives and factors?

  • If uncertain → Use Blindspot Detector prompt
  • If confident → Continue to step 5

05

Have I imagined how this decision could fail?

  • If no → Use Premortem Facilitator prompt
  • If yes → You're ready to make an informed decision

15 of 23

Real-World Application

Case Scenario: Marketing Budget Allocation Decision

Context:

100-person B2B SaaS company, must allocate £500K marketing budget between paid ads, content marketing, and event sponsorships

Traditional Approach:

Research industry benchmarks, see what similar companies spend, follow "best practice" of 40% paid, 40% content, 20% events

16 of 23

Critical Analysis using Prompts:

Assumption Challenger Results:

Critical assumption: "Our ICP responds to content marketing like industry average"

Confidence: 4/10 (low confidence)

Test: £5K content experiment with current leads

Impact if wrong: Wasted £200K on ineffective content

Contrarian Analysis Results:

AI argued against content marketing focus, citing long sales cycles, complex B2B buying process, and competitor content saturation. Suggested events might provide better direct pipeline impact.

17 of 23

Blindspot Detection Results:

Identified overlooked factors: customer acquisition cost by channel, seasonal buying patterns, sales team capacity to handle different lead types, and economic uncertainty affecting event attendance.

Premortem Results:

Predicted failure through spreading budget too thin across channels, lack of clear attribution tracking, and sales team confusion about lead quality differences between channels.

Critical Analysis Allocation:

60% paid ads (highest certainty), 25% events (high-value prospects), 15% content (test and optimise)

Results:

2.3x better ROI than industry benchmark allocation, clearer attribution, more qualified pipeline.

Key Insight: The "optimal" allocation looked wrong by industry standards but was right for their specific context and constraints.

18 of 23

Common Mistakes

What to Avoid

Seeking AI validation

Focus on getting challenged instead

Using leading questions

Frame questions neutrally or adversarially

Accepting first response

Push AI to be more critical if it seems agreeable

Ignoring uncomfortable feedback

The discomfort is where insights live

19 of 23

What to Do Instead

Start with "convince me I'm wrong"

Force AI into analytical mode

Add context about your biases

Help AI understand what to challenge

Ask for specific evidence

Don't accept general disagreement

Request failure scenarios

Before making commitments, imagine how they fail

20 of 23

Weekly Practice Implementation Framework

1

Monday: Assumption Audit

List all major assumptions in this week's decisions

Use Assumption Challenger on one key decision

Identify which assumptions need testing vs. which are actually facts

2

Tuesday: Devil's Advocate Day

Choose one decision where team consensus feels too easy

Use Contrarian Analyst to argue against the popular position

Document new risks and downsides identified

21 of 23

1

Wednesday: Blindspot Wednesday

Pick one complex decision involving multiple stakeholders

Use Blindspot Detector to reveal missing perspectives

List questions you should have been asking but weren't

2

Thursday: Failure Thursday

Select one major upcoming decision or commitment

Use Premortem Facilitator to predict realistic failure scenarios

Build early warning systems and preventive measures

22 of 23

1

Friday: Review and Reflect

Compare decision quality this week vs. previous weeks

Identify patterns in your assumption-making and blindspots

Plan which critical analysis techniques to apply next week

Monthly Deep Dive:

  • Choose one major decision from the month
  • Run through complete 4-prompt sequence
  • Compare results to decisions made without critical analysis
  • Update team processes for using AI more strategically

23 of 23

Notes & Customisation

Team-Specific Adaptations

For Leadership Teams:

Focus prompts on strategic assumptions and long-term blindspots

Emphasise stakeholder perspectives and competitive dynamics

Include succession planning and organisational capability constraints

For Product Teams:

Adapt prompts for feature prioritisation and user assumption testing

Include technical debt vs. new feature trade-offs

Focus on user behaviour assumptions and market timing

For Operations Teams:

Customise for process improvement and automation decisions

Include operational risk assessment and change management blindspots

Focus on scalability assumptions and implementation complexity

©2025 AI Prompt Hackers. All Rights Reserved