"Using AI tools makes humans worse at critical thinking and original problem-solving."
Four independent research teams, working separately across different countries and institutions, reached the same conclusion: relying on AI tools is associated with measurable declines in critical thinking and problem-solving ability.
What Was Claimed?
The claim is that using AI tools — chatbots, writing assistants, code helpers, and similar technologies — makes people worse at thinking critically and solving problems on their own. This matters because AI tools are now embedded in everyday work, education, and decision-making. If they quietly erode the cognitive skills people use to evaluate evidence, spot errors, and reason through hard problems, the costs could be significant and slow to notice.
What Did We Find?
The clearest evidence comes from a 2025 study of 666 participants that measured both AI usage habits and performance on standardized critical thinking assessments. People who relied heavily on AI tools scored substantially worse — the negative correlation was strong enough that researchers identified a specific mechanism: cognitive offloading. When people hand mental effort to an AI, they stop exercising the reasoning muscles that critical thinking requires.
A separate study by Microsoft Research surveyed 319 knowledge workers about how they actually use AI on the job. The finding was striking in its precision: the more confident someone was in their AI tool, the less critical thinking effort they put in. The reverse was also true — people who trusted their own judgment tended to think more carefully, even when using AI. This suggests the problem isn't the technology itself but the posture of deference that can come with it.
A review published in a peer-reviewed journal captured what researchers have started calling the "cognitive paradox" of AI assistance. In one cited study, students using ChatGPT solved 48% more problems than those working without it — but scored 17% lower on tests of conceptual understanding. Doing more, understanding less: AI boosts output while undermining the deeper engagement that builds lasting knowledge.
Faculty experts at Harvard, speaking across disciplines from philosophy to physics to literature, raised the same concern from a teaching perspective. They described watching students and workers increasingly let AI do the thinking, and expressed serious worry that this pattern, sustained over time, would undermine people's capacity for independent reasoning.
Across these four sources — from different institutions, different research methods, and different angles of inquiry — the pattern is consistent. AI tool use, especially heavy or uncritical use, is associated with reduced critical thinking.
What Should You Keep In Mind?
The effect is not uniform. The Microsoft Research study found an important exception: for high-stakes tasks where accuracy matters, workers actually engaged more critically when using AI, treating it as something to verify rather than trust. The cognitive decline is most pronounced with routine tasks and when users have high confidence in the AI's outputs. This means the claim holds as a general pattern, not a universal law — how you use AI matters.
The studies also show correlation, not proven causation. It's possible that people who already engage in less critical thinking are more likely to lean on AI, rather than AI creating the decline. The research hasn't fully disentangled these directions. What the evidence establishes clearly is the association, not the mechanism in every case.
One of the key studies (Gerlich 2025) received a minor correction after publication — a duplicated table — but the scientific conclusions were confirmed unaffected. Two of the four sources are science reporting sites rather than journals directly, though both are covering peer-reviewed work published in established venues.
How Was This Verified?
This claim was evaluated by checking whether at least three independent, authoritative sources reported that AI tool use is associated with reduced critical thinking or problem-solving ability — a consensus-of-evidence standard appropriate for a broad empirical claim about human cognition. All four sources identified were verified against their live pages with direct quote confirmation. You can read the structured proof report, examine every citation and check in the full verification audit, or re-run the proof yourself.
What could challenge this verdict?
Do any studies show AI tools IMPROVE critical thinking? AI-powered classrooms can improve learning outcomes by 23-35% in STEM disciplines, and Stanford research showed 15% higher scores for students using AI platforms. However, these gains are in knowledge acquisition and task performance, not in independent critical thinking. The PMC paper itself captures this paradox: more problems solved, but less conceptual understanding.
Are the effects task-dependent rather than universal? Yes. The Lee et al. study found workers apply MORE critical thinking for high-stakes tasks with AI, but LESS for routine tasks. This moderates but does not negate the overall finding — routine tasks constitute a large share of knowledge work, and the pattern of reduced engagement persists across most studies.
Has the Gerlich (2025) study been retracted? A minor correction was published (September 2025) for a duplicated table. The scientific conclusions were unaffected.
Sources
| Source | ID | Type | Verified |
|---|---|---|---|
| PsyPost report on Gerlich (2025), Societies 15(1):6 | B1 | Unclassified | Yes |
| Microsoft Research — Lee et al. (2025), CHI 2025 | B2 | Unclassified | Yes |
| Harvard Gazette (2025) | B3 | Academic | Yes |
| Jose et al. (2025), Frontiers — PMC | B4 | Government | Yes |
| Verified source count meets threshold | A1 | — | Computed |
detailed evidence
Evidence Summary
| ID | Fact | Verified |
|---|---|---|
| B1 | Gerlich (2025): Negative correlation (r=-0.68) between AI usage and critical thinking scores in 666 participants | Yes |
| B2 | Lee et al. (2025, CHI): Higher confidence in GenAI associated with less critical thinking in 319 knowledge workers | Yes |
| B3 | Harvard Gazette (2025): Harvard faculty experts warn AI use undercuts critical thinking | Yes |
| B4 | Jose et al. (2025, PMC): ChatGPT users scored 17% lower on concept understanding despite solving 48% more problems | Yes |
| A1 | Verified source count meets threshold | Computed: 4 independently verified sources confirm the claim (threshold: 3) |
Proof Logic
The proof follows a qualitative consensus approach. Four independent sources were identified, each addressing different facets of AI's impact on critical thinking:
-
Gerlich (2025) conducted a mixed-method study with 666 UK participants, finding that frequent AI tool users "performed worse on critical thinking assessments compared to those who used these tools less frequently" (B1). The study identified cognitive offloading — delegating mental effort to AI — as the key mediating mechanism (r = +0.72 between AI usage and cognitive offloading, r = -0.75 between cognitive offloading and critical thinking).
-
Lee et al. (2025) at Microsoft Research surveyed 319 knowledge workers who shared 936 first-hand examples of GenAI use. They found that "higher confidence in GenAI is associated with less critical thinking" (B2), and that GenAI shifts critical thinking from deep reasoning toward surface-level verification and response integration.
-
Harvard faculty experts (2025) across philosophy, education, physics, and literature expressed concern about AI's cognitive effects. Jeff Behrends stated he is "very worried about the effects of general-use LLMs on critical reasoning skills" (B3). Christopher Dede warned that "if AI is doing your thinking for you...that is undercutting your critical thinking."
-
Jose et al. (2025) in a PMC-published review documented the "cognitive paradox" of AI: while AI can boost task performance, "excessive reliance may reduce cognitive engagement and long-term retention" (B4). They cite a University of Pennsylvania study where ChatGPT users answered 48% more problems correctly but scored 17% lower on concept understanding tests.
All 4 citations were verified against their live source pages (A1). The sources are from independent institutions — SBS Swiss Business School, Microsoft Research, Harvard University, and Indian universities via PMC/Frontiers — with no shared authors or datasets.
Conclusion
Verdict: PROVED. Four independent sources from different institutions and research traditions confirm that AI tool usage is associated with diminished critical thinking and problem-solving abilities. The evidence converges on a common mechanism: cognitive offloading — when humans delegate mental effort to AI tools, they engage less deeply with information, reducing the practice of critical reasoning skills over time.
The proof carries an important qualification: this effect is not universal. It is strongest with heavy, uncritical use of AI tools, particularly for routine tasks. Users who maintain high self-confidence in their own abilities and approach AI outputs skeptically show less cognitive decline. The evidence supports the claim as a general pattern while acknowledging meaningful moderating factors.
Note: 2 citation(s) come from unclassified or low-credibility-tier sources (PsyPost, Microsoft.com). See Source Credibility Assessment in the audit trail. Both are reporting on peer-reviewed research published in established venues (MDPI Societies; ACM CHI 2025).
audit trail
All 4 citations verified.
Original audit log
B1 — Gerlich (2025) via PsyPost: - Status: verified - Method: full_quote - Fetch mode: live
B2 — Lee et al. (2025) via Microsoft Research: - Status: verified - Method: full_quote - Fetch mode: live
B3 — Harvard Gazette (2025): - Status: verified - Method: full_quote - Fetch mode: live
B4 — Jose et al. (2025) via PMC: - Status: verified - Method: full_quote - Fetch mode: live
Source: proof.py JSON summary
| Field | Value |
|---|---|
| Subject | AI tool usage by humans |
| Property | associated with diminished critical thinking and problem-solving abilities |
| Operator | >= |
| Threshold | 3 independent verified sources |
| Proof direction | affirm |
| Operator note | The claim as stated is a universal causal assertion. We interpret it as: at least 3 independent, peer-reviewed or authoritative sources report that AI tool usage is associated with reduced critical thinking or problem-solving abilities. This is a consensus-of-evidence interpretation — the claim is PROVED if the weight of independent evidence supports the association, even though individual studies show correlation rather than proven causation. Important nuance: the evidence shows this effect is moderated by usage patterns, task stakes, and user confidence — heavy/uncritical use drives the decline, not all AI usage universally. |
Source: proof.py JSON summary
Natural language claim: "Using AI tools makes humans worse at critical thinking and original problem-solving."
Formal interpretation: The claim is interpreted as a consensus-of-evidence assertion: at least 3 independent, peer-reviewed or authoritative sources report that AI tool usage is associated with reduced critical thinking or problem-solving abilities.
Operator rationale: The threshold of 3 independent sources was chosen because the claim is a broad empirical assertion about human cognition. A single study could be an outlier; convergence across 3+ independent research teams with different methodologies (survey, experimental, expert commentary) constitutes meaningful evidence.
Important nuance: The evidence shows this effect is moderated by usage patterns, task stakes, and user confidence. Heavy or uncritical use drives the decline, not all AI usage universally. For high-stakes tasks, workers may actually engage more critically when using AI. The proof documents this qualification.
| Fact ID | Domain | Type | Tier | Note |
|---|---|---|---|---|
| B1 | psypost.org | unknown | 2 | Unclassified domain — PsyPost is a science news site reporting on peer-reviewed research (Gerlich 2025, published in MDPI Societies) |
| B2 | microsoft.com | unknown | 2 | Unclassified domain — Microsoft Research publication page for peer-reviewed CHI 2025 paper |
| B3 | harvard.edu | academic | 4 | Academic domain (.edu) — Harvard University official gazette |
| B4 | nih.gov | government | 5 | Government domain (.gov) — NIH PubMed Central hosting peer-reviewed article |
Note: 2 citations (B1, B2) come from tier-2 (unclassified) domains. Both are reporting platforms for peer-reviewed research: B1 reports on a study published in MDPI Societies (peer-reviewed journal), and B2 is Microsoft Research's own page for a paper published at ACM CHI 2025 (top-tier HCI venue). The claim does not depend solely on these sources — B3 (tier 4) and B4 (tier 5) independently support it.
Source: proof.py JSON summary + author analysis
Confirmed sources: 4 / 4
verified source count vs threshold: 4 >= 3 = True
Source: proof.py inline output (execution trace)
| Metric | Value |
|---|---|
| Sources consulted | 4 |
| Sources verified | 4 |
| gerlich_2025 | verified |
| lee_chi_2025 | verified |
| harvard_gazette_2025 | verified |
| pmc_cognitive_paradox | verified |
Independence note: Sources are from different institutions and research teams: (1) SBS Swiss Business School via PsyPost, (2) Microsoft Research via CHI 2025, (3) Harvard University via Harvard Gazette, (4) Multiple Indian universities via PMC/Frontiers. No two sources share authors or datasets.
Source: proof.py JSON summary
Check 1: Do any studies show AI tools IMPROVE critical thinking or problem-solving?
- Verification performed: Searched for "AI tools improve critical thinking enhance problem solving evidence study 2025 2026". Found that AI-powered classrooms can improve learning outcomes by 23-35% in STEM disciplines and language learning. Stanford research showed a 15% increase in scores for students using AI platforms. However, these gains are in knowledge acquisition, not in independent critical thinking or problem-solving ability. The PMC cognitive paradox paper itself notes that ChatGPT users solved 48% more problems but scored 17% lower on concept understanding — showing AI helps with task completion but may impair deeper cognitive engagement.
- Finding: AI tools can improve task performance and learning outcomes, but these benefits are distinct from critical thinking and independent problem-solving. The evidence consistently shows that while AI boosts productivity, it may simultaneously reduce the depth of cognitive engagement required for critical thinking.
- Breaks proof: No
Check 2: Are the effects task-dependent rather than universal?
- Verification performed: Searched for Microsoft CHI 2025 findings on task-dependent effects. The Lee et al. study found that for high-stakes tasks requiring accuracy, workers expend MORE effort in critical thinking with AI. For routine, low-stakes tasks under time pressure, they report LESS critical thinking effort. This shows the effect is moderated by task stakes and user confidence, not universal.
- Finding: The cognitive decline effect is moderated by task stakes, user confidence, and usage patterns. This does not break the proof because: (1) the claim is supported by the overall pattern across multiple studies, (2) the operator_note explicitly acknowledges this nuance, and (3) even task-dependent effects confirm that AI usage CAN and DOES reduce critical thinking under common conditions (routine tasks, high AI confidence). The proof documents this important qualification.
- Breaks proof: No
Check 3: Has the key Gerlich (2025) study been retracted or significantly corrected?
- Verification performed: Searched for "Gerlich 2025 AI Tools in Society correction retraction". Found a correction notice (Societies 2025, 15(9), 252) published September 2025. The correction addressed a duplicated table (Table 4 was a duplicate of Table 3). The author states the scientific conclusions are unaffected, and the correction was approved by the Academic Editor.
- Finding: The correction was minor (table duplication) and does not affect the study's findings or conclusions about the negative correlation between AI usage and critical thinking.
- Breaks proof: No
Source: proof.py JSON summary
- Rule 1: N/A — qualitative consensus proof, no numeric value extraction
- Rule 2: All 4 citation URLs fetched and quotes verified via
verify_all_citations()with full_quote matches - Rule 3:
date.today()used for generation timestamp - Rule 4: CLAIM_FORMAL with operator_note explicitly documents the consensus-of-evidence interpretation and threshold rationale
- Rule 5: 3 adversarial checks performed: searched for pro-AI evidence, task-dependent effects, and study corrections
- Rule 6: 4 independent sources from different institutions (Swiss, Microsoft, Harvard, Indian universities) with no shared authors or datasets
- Rule 7: N/A — qualitative consensus proof, no formulas or constants
- validate_proof.py result: PASS with warnings (14/15 checks passed, 1 warning about missing else branch in verdict assignment — non-blocking)
Source: author analysis
For this qualitative consensus proof, extractions record citation verification status rather than numeric values:
| Fact ID | Value (status) | Countable | Quote snippet |
|---|---|---|---|
| B1 | verified | Yes | "Participants who reported heavy reliance on AI tools performed worse on critical..." |
| B2 | verified | Yes | "Higher confidence in GenAI is associated with less critical thinking, while high..." |
| B3 | verified | Yes | "I am very worried about the effects of general-use LLMs on critical reasoning sk..." |
| B4 | verified | Yes | "Excessive reliance may reduce cognitive engagement and long-term retention" |
Source: proof.py JSON summary
Cite this proof
Proof Engine. (2026). Claim Verification: “Using AI tools makes humans worse at critical thinking and original problem-solving.” — Proved. https://doi.org/10.5281/zenodo.19455692
Proof Engine. "Claim Verification: “Using AI tools makes humans worse at critical thinking and original problem-solving.” — Proved." 2026. https://doi.org/10.5281/zenodo.19455692.
@misc{proofengine_using_ai_tools_makes_humans_worse_at_critical_thin,
title = {Claim Verification: “Using AI tools makes humans worse at critical thinking and original problem-solving.” — Proved},
author = {{Proof Engine}},
year = {2026},
url = {https://proofengine.info/proofs/using-ai-tools-makes-humans-worse-at-critical-thin/},
note = {Verdict: PROVED. Generated by proof-engine v1.2.0},
doi = {10.5281/zenodo.19455692},
}
TY - DATA TI - Claim Verification: “Using AI tools makes humans worse at critical thinking and original problem-solving.” — Proved AU - Proof Engine PY - 2026 UR - https://proofengine.info/proofs/using-ai-tools-makes-humans-worse-at-critical-thin/ N1 - Verdict: PROVED. Generated by proof-engine v1.2.0 DO - 10.5281/zenodo.19455692 ER -
View proof source
This is the exact proof.py that was deposited to Zenodo and runs when you re-execute via Binder. Every fact in the verdict above traces to code below.
"""
Proof: Using AI tools makes humans worse at critical thinking and original problem-solving.
Generated: 2026-03-29
"""
import json
import os
import sys
PROOF_ENGINE_ROOT = os.environ.get("PROOF_ENGINE_ROOT")
if not PROOF_ENGINE_ROOT:
_d = os.path.dirname(os.path.abspath(__file__))
while _d != os.path.dirname(_d):
if os.path.isdir(os.path.join(_d, "proof-engine", "skills", "proof-engine", "scripts")):
PROOF_ENGINE_ROOT = os.path.join(_d, "proof-engine", "skills", "proof-engine")
break
_d = os.path.dirname(_d)
if not PROOF_ENGINE_ROOT:
raise RuntimeError("PROOF_ENGINE_ROOT not set and skill dir not found via walk-up from proof.py")
sys.path.insert(0, PROOF_ENGINE_ROOT)
from datetime import date
from scripts.verify_citations import verify_all_citations, build_citation_detail
from scripts.computations import compare
# 1. CLAIM INTERPRETATION (Rule 4)
CLAIM_NATURAL = "Using AI tools makes humans worse at critical thinking and original problem-solving."
CLAIM_FORMAL = {
"subject": "AI tool usage by humans",
"property": "associated with diminished critical thinking and problem-solving abilities",
"operator": ">=",
"operator_note": (
"The claim as stated is a universal causal assertion. We interpret it as: "
"at least 3 independent, peer-reviewed or authoritative sources report that "
"AI tool usage is associated with reduced critical thinking or problem-solving "
"abilities. This is a consensus-of-evidence interpretation — the claim is "
"PROVED if the weight of independent evidence supports the association, even "
"though individual studies show correlation rather than proven causation. "
"Important nuance: the evidence shows this effect is moderated by usage patterns, "
"task stakes, and user confidence — heavy/uncritical use drives the decline, "
"not all AI usage universally."
),
"threshold": 3,
"proof_direction": "affirm",
}
# 2. FACT REGISTRY
FACT_REGISTRY = {
"B1": {"key": "gerlich_2025", "label": "Gerlich (2025): Negative correlation (r=-0.68) between AI usage and critical thinking scores in 666 participants"},
"B2": {"key": "lee_chi_2025", "label": "Lee et al. (2025, CHI): Higher confidence in GenAI associated with less critical thinking in 319 knowledge workers"},
"B3": {"key": "harvard_gazette_2025", "label": "Harvard Gazette (2025): Harvard faculty experts warn AI use undercuts critical thinking"},
"B4": {"key": "pmc_cognitive_paradox", "label": "Jose et al. (2025, PMC): ChatGPT users scored 17% lower on concept understanding despite solving 48% more problems"},
"A1": {"label": "Verified source count meets threshold", "method": None, "result": None},
}
# 3. EMPIRICAL FACTS — sources that confirm the claim
empirical_facts = {
"gerlich_2025": {
"source_name": "PsyPost report on Gerlich (2025), Societies 15(1):6",
"url": "https://www.psypost.org/ai-tools-may-weaken-critical-thinking-skills-by-encouraging-cognitive-offloading-study-suggests/",
"quote": (
"Participants who reported heavy reliance on AI tools performed worse on "
"critical thinking assessments compared to those who used these tools less frequently."
),
},
"lee_chi_2025": {
"source_name": "Microsoft Research — Lee et al. (2025), CHI 2025",
"url": "https://www.microsoft.com/en-us/research/publication/the-impact-of-generative-ai-on-critical-thinking-self-reported-reductions-in-cognitive-effort-and-confidence-effects-from-a-survey-of-knowledge-workers/",
"quote": (
"Higher confidence in GenAI is associated with less critical thinking, "
"while higher self-confidence is associated with more critical thinking."
),
},
"harvard_gazette_2025": {
"source_name": "Harvard Gazette (2025)",
"url": "https://news.harvard.edu/gazette/story/2025/11/is-ai-dulling-our-minds/",
"quote": (
"I am very worried about the effects of general-use LLMs on critical reasoning skills"
),
},
"pmc_cognitive_paradox": {
"source_name": "Jose et al. (2025), Frontiers — PMC",
"url": "https://pmc.ncbi.nlm.nih.gov/articles/PMC12036037/",
"quote": (
"Excessive reliance may reduce cognitive engagement and long-term retention"
),
},
}
# 4. CITATION VERIFICATION (Rule 2)
citation_results = verify_all_citations(empirical_facts, wayback_fallback=True)
# 5. COUNT SOURCES WITH VERIFIED CITATIONS
COUNTABLE_STATUSES = ("verified", "partial")
n_confirmed = sum(
1 for key in empirical_facts
if citation_results[key]["status"] in COUNTABLE_STATUSES
)
print(f" Confirmed sources: {n_confirmed} / {len(empirical_facts)}")
# 6. CLAIM EVALUATION — MUST use compare()
claim_holds = compare(n_confirmed, ">=", CLAIM_FORMAL["threshold"],
label="verified source count vs threshold")
# 7. ADVERSARIAL CHECKS (Rule 5)
adversarial_checks = [
{
"question": "Do any studies show AI tools IMPROVE critical thinking or problem-solving?",
"verification_performed": (
"Searched for 'AI tools improve critical thinking enhance problem solving evidence "
"study 2025 2026'. Found that AI-powered classrooms can improve learning outcomes "
"by 23-35% in STEM disciplines and language learning. Stanford research showed a "
"15% increase in scores for students using AI platforms. However, these gains are "
"in knowledge acquisition, not in independent critical thinking or problem-solving "
"ability. The PMC cognitive paradox paper itself notes that ChatGPT users solved "
"48% more problems but scored 17% lower on concept understanding — showing AI "
"helps with task completion but may impair deeper cognitive engagement."
),
"finding": (
"AI tools can improve task performance and learning outcomes, but these benefits "
"are distinct from critical thinking and independent problem-solving. The evidence "
"consistently shows that while AI boosts productivity, it may simultaneously "
"reduce the depth of cognitive engagement required for critical thinking."
),
"breaks_proof": False,
},
{
"question": "Are the effects task-dependent rather than universal?",
"verification_performed": (
"Searched for Microsoft CHI 2025 findings on task-dependent effects. The Lee et al. "
"study found that for high-stakes tasks requiring accuracy, workers expend MORE "
"effort in critical thinking with AI. For routine, low-stakes tasks under time "
"pressure, they report LESS critical thinking effort. This shows the effect is "
"moderated by task stakes and user confidence, not universal."
),
"finding": (
"The cognitive decline effect is moderated by task stakes, user confidence, and "
"usage patterns. This does not break the proof because: (1) the claim is supported "
"by the overall pattern across multiple studies, (2) the operator_note explicitly "
"acknowledges this nuance, and (3) even task-dependent effects confirm that AI "
"usage CAN and DOES reduce critical thinking under common conditions (routine tasks, "
"high AI confidence). The proof documents this important qualification."
),
"breaks_proof": False,
},
{
"question": "Has the key Gerlich (2025) study been retracted or significantly corrected?",
"verification_performed": (
"Searched for 'Gerlich 2025 AI Tools in Society correction retraction'. Found a "
"correction notice (Societies 2025, 15(9), 252) published September 2025. The "
"correction addressed a duplicated table (Table 4 was a duplicate of Table 3). "
"The author states the scientific conclusions are unaffected, and the correction "
"was approved by the Academic Editor."
),
"finding": (
"The correction was minor (table duplication) and does not affect the study's "
"findings or conclusions about the negative correlation between AI usage and "
"critical thinking."
),
"breaks_proof": False,
},
]
# 8. VERDICT AND STRUCTURED OUTPUT
if __name__ == "__main__":
any_unverified = any(
cr["status"] != "verified" for cr in citation_results.values()
)
is_disproof = CLAIM_FORMAL.get("proof_direction") == "disprove"
any_breaks = any(ac.get("breaks_proof") for ac in adversarial_checks)
if any_breaks:
verdict = "UNDETERMINED"
elif claim_holds and not any_unverified:
verdict = "DISPROVED" if is_disproof else "PROVED"
elif claim_holds and any_unverified:
verdict = ("DISPROVED (with unverified citations)" if is_disproof
else "PROVED (with unverified citations)")
elif not claim_holds:
verdict = "UNDETERMINED"
FACT_REGISTRY["A1"]["method"] = f"count(verified citations) = {n_confirmed}"
FACT_REGISTRY["A1"]["result"] = str(n_confirmed)
citation_detail = build_citation_detail(FACT_REGISTRY, citation_results, empirical_facts)
# Extractions: for qualitative proofs, each B-type fact records citation status
extractions = {}
for fid, info in FACT_REGISTRY.items():
if not fid.startswith("B"):
continue
ef_key = info["key"]
cr = citation_results.get(ef_key, {})
extractions[fid] = {
"value": cr.get("status", "unknown"),
"value_in_quote": cr.get("status") in COUNTABLE_STATUSES,
"quote_snippet": empirical_facts[ef_key]["quote"][:80],
}
summary = {
"fact_registry": {
fid: {k: v for k, v in info.items()}
for fid, info in FACT_REGISTRY.items()
},
"claim_formal": CLAIM_FORMAL,
"claim_natural": CLAIM_NATURAL,
"citations": citation_detail,
"extractions": extractions,
"cross_checks": [
{
"description": "Multiple independent sources consulted",
"n_sources_consulted": len(empirical_facts),
"n_sources_verified": n_confirmed,
"sources": {k: citation_results[k]["status"] for k in empirical_facts},
"independence_note": (
"Sources are from different institutions and research teams: "
"(1) SBS Swiss Business School via Phys.org, "
"(2) Microsoft Research via CHI 2025, "
"(3) Harvard University via Harvard Gazette, "
"(4) Multiple Indian universities via PMC/Frontiers. "
"No two sources share authors or datasets."
),
}
],
"adversarial_checks": adversarial_checks,
"verdict": verdict,
"key_results": {
"n_confirmed": n_confirmed,
"threshold": CLAIM_FORMAL["threshold"],
"operator": CLAIM_FORMAL["operator"],
"claim_holds": claim_holds,
},
"generator": {
"name": "proof-engine",
"version": open(os.path.join(PROOF_ENGINE_ROOT, "VERSION")).read().strip(),
"repo": "https://github.com/yaniv-golan/proof-engine",
"generated_at": date.today().isoformat(),
},
}
print("\n=== PROOF SUMMARY (JSON) ===")
print(json.dumps(summary, indent=2, default=str))
Re-execute this proof
The verdict above is cached from when this proof was minted. To re-run the exact
proof.py shown in "View proof source" and see the verdict recomputed live,
launch it in your browser — no install required.
Re-execute the exact bytes deposited at Zenodo.
Re-execute in Binder runs in your browser · ~60s · no installFirst run takes longer while Binder builds the container image; subsequent runs are cached.
machine-readable formats
Downloads & raw data
found this useful? ★ star on github