ARTICLES BIAIS ET ETHIQUES IA

AI confirmation bias. Question rewrites for ethical prompting

alt="image décorative IA"
AI Confirmation Bias: Question Rewrites for Ethical Prompting

AI Confirmation Bias: Question Rewrites for Ethical Prompting

Introduction: The Hidden Cost of Leading Questions

When your question hints at the answer you want, AI usually gives it to you.

You will get reasons it works. You will not get where it fails, what to try instead, or what research says about limits. The first question shapes the frame. The frame shapes the answer.

This article shows you how to spot the most common traps and gives you two practical rewrites you can use today. I use an ethical prompting framework that’s specific for therapists, writers, and organisations. It focuses on professional ethics, not just accuracy.

Critical Connection: In high-stakes fields like healthcare, the way a question is asked can shape triage, prioritisation, and even who receives urgent attention first. That’s why medical prompting isn’t just about accuracy—it’s about ethics, bias, and fair access to care. For more on this, see our article on medical prompts in diagnosis and access to care.

Understanding Confirmation Bias in Plain Language

Confirmation bias means we look for information that supports what we already think. Everyone does it. With AI tools, this trend grows.

These systems are designed to be helpful and follow your lead. If your question assumes something is true, the model often stays inside that frame. Research backs this. Studies on sycophancy show that helpful systems echo user beliefs. Framing studies show that positive versus negative wording shifts answers. Persona studies show that assigned roles change stance and tone.

The Two-Minute Test: See Bias in Action

None of this is theory. You can see it in two minutes:

Step 1: Start a new chat. Ask, « Why is remote work better for productivity? »

Step 2: Start another new chat. Ask, « Compare remote work and office work for productivity. When does each work better? What do local labour rules and norms change? »

Step 3: Compare the two answers.

The first collects reasons that support your claim. The second gives conditions, trade-offs, and local considerations. Same tool. Same topic. Different frame. Different quality.

Five Question Patterns That Create Bias

Pattern 1: Asking « Why » When You Should Ask « When »

You might ask: « Why is my therapy approach effective for anxiety? »

Problem: « Why » assumes it works. The answer explains why, not whether.

Better: « When does this approach work for anxiety? When does it fall short? What does research say about limits? »

Why this matters for therapists: Clinical work needs proper conditions and safeguards, not just success stories.

Pattern 2: Giving Only Two Choices

You might wonder: « Should we pick tool A or tool B for grant management? »

The issue: You limit the choice to just those two options, and the model will reflect that.

Better: « What are the various ways a small nonprofit in France can manage grants? Include options we might not know about. For each, list when it works best and when it fails. »

Why this matters for organisations: Budgets, data protection, and maintenance costs vary based on the context. Binary questions hide those trade-offs.

Pattern 3: Telling AI to Be Nice

You might ask: « Act as a supportive coach and review my writing. »

The issue: A supportive tone can lead to excessive praise.

Better: « Review this chapter from two perspectives. First, what works. Second, what would make an agent reject it today. Be specific about both. »

Why this matters for writers: Publishing values clear critique over mere encouragement.

Pattern 4: Long Setup That Assumes You Are Right

You might ask: « I have been using technique X with great results and clients love it. How should I teach it to others? »

Problem: You shared your technique, so others will likely accept that view.

Better: « I use technique X. What assumptions am I making? Where might this not fit? What checks should I run before teaching it? »

Why this matters for therapists and trainers: You need to test assumptions before sharing a method.

Pattern 5: Only Asking About Success

You might ask: « How do I launch our new service successfully in Belgium? »

The problem: You only think about success. Failure analysis reveals more.

Better: « Imagine our service launch in Belgium failed after six months. What are five likely reasons? For each, what early warning sign should I watch for? »

Why this matters for organisations: Pre-mortem thinking surfaces risks before you commit resources. In Europe, market entry costs are high, and corrections are expensive.

Two Templates You Can Use Today

These two reduce confirmation bias across tools. I provide a tailored framework that centres on ethical checks, boundary language, and governance steps.

Template 1: Force Both Sides

« What are the benefits and drawbacks of [your approach]? For each drawback, suggest one way to test whether it applies to my situation. »

Therapist’s Example:
« What are the pros and cons of narrative therapy for complex trauma and dissociative features? For each drawback, suggest one safety check I can run with clinical supervision. »

Template 2: Request the Opposite View

« Make the strongest case against [your position]. Be specific about costs, risks, and failure scenarios. »

Nonprofit Example:
« List the main drawbacks of this donor management software for a small nonprofit. Include training time, data export limits, and data protection gaps. »

For the complete set: The client framework adds ethics verification, professional boundary checks, and compliance prompts that fit your jurisdiction and sector.

Why Professional Ethics Need More Than Generic Templates

For Therapists

  • Safeguards for client welfare
  • Referral triggers
  • Crisis redirection language
  • Supervision alignment

For Writers

  • Protect your voice
  • Keep critique and direction separate
  • Check for originality
  • Verify suggestions with your skills

For Non-profits and Teams

  • Align missions
  • Ensure the duty of care
  • Test budget realities
  • Check privacy compliance with regulations like GDPR in the EU, CCPA in California, and local sector rules

Generic bias reduction isn’t enough if your work impacts well-being, careers, budgets, or public trust.

What Research Shows

Sycophancy: Helpful assistants usually agree with the user, but this can harm accuracy.

Framing Sensitivity: Positive versus negative wording shifts answers in measurable ways.

Persona Effects: Assigned roles, like supportive coach or critical editor, change the stance and tone.

Context Memory: Past messages in this thread can affect how later answers are given.

Independent studies confirm these behavioural patterns. Anthropic researchers (2023) identified sycophancy—models echoing user beliefs. ACL 2024 papers described how persona and framing affect stance. Major model documentation also explains that chat history shapes later replies.

When to Consider Professional Guidance

In clinical or regulated environments—hospitals, pharma, public health—you also need to consider structural bias in AI models themselves: under-representation in training data, unequal access to tools, language gaps, and legal accountability. We explore that side in depth in our guide to medical prompts, bias, and fair access in healthcare.

Use the Templates For:

  • Everyday questions
  • General professional work
  • Initial exploration

Bring in a Field-Aware Framework When You Face:

  • Clinical decisions where mistakes affect client welfare
  • Publication-level writing where you need adversarial feedback, not validation
  • Organizational commitments with real money, data, or long contracts
  • Any decision where professional ethics matter as much as accuracy
  • Team-wide adoption where many people need consistent practice

Your Quick Self-Check

Before you trust any AI answer on a professional decision, check these:

  • Did I avoid « why » questions that assume truth?
  • Did I ask what could go wrong, not only what could go right?
  • Did I request at least one opposing view?
  • Did I consider the ethical implications for my field and role?
  • Do I have a plan to verify one claim in the real world this week?

If I checked fewer than four, should I rewrite my question first?

Frequently Asked Questions

Does this work across different AI tools?
Yes. The effects come from how modern assistants are trained and how they carry context. The two chat tests will show the pattern in any mainstream tool.
I am not technical. Is this too complex?
No. You need to spot the patterns and use the two templates. Technical knowledge is optional.
What about health-related questions?
Never use AI to diagnose or treat. Use it to challenge your thinking and surface research. Follow local laws and professional rules in your practice. If safety is at risk, direct clients to local crisis services.
What about compliance and data protection?
Compliance checks are key to ethical prompting for organisations. They help ensure privacy and safety standards are met. Check for likely obligations in your area. Then, confirm with your privacy lead or legal counsel.
Can I share these templates with my team?
Yes. Many teams print the two templates for reference. For team governance and training, plan a structured rollout. Also, include measurement and regular reviews.

Your Next Step

Step 1: Test – Run the two chat tests. Pick one important question.

Step 2: Rewrite – Rewrite it using Template 1 or Template 2.

Step 3: Notice – Notice how a small change in framing shifts the quality, fairness, and reliability of the answer.

These micro-adjustments are where ethical prompting begins.

Ready to Apply This Within Your Field?

If you’re ready to apply this within your field, book an initial consultation to discuss your goals and context. This meeting defines your needs and the scope of work.

From there, we can design your field-specific ethical prompting framework—integrating bias control, regulatory alignment, and compliance safeguards.

Research Methodology & Transparency

Research Phase: This guide synthesizes insights from academic research on AI behavior, professional ethics frameworks, and direct experience consulting with healthcare practitioners, creative professionals, and nonprofit organizations across multiple jurisdictions.

AI Tool Usage: This article was developed using AI tools as writing and research assistants. Claude assisted with content structuring and initial drafting. ChatGPT supported research synthesis. All final analysis, ethical frameworks, and professional recommendations reflect my expertise in AI ethics consulting.

Source Verification: All research findings and regulatory references were verified against original publications as of January 2025.

Sources and References

Research Studies

Regulatory Frameworks & Ethics

Related Articles

This article was co-written with AI assistance. Document created: January 8, 2025 | For: Dieneba LESDEMA – Prompt & Pulse