πŸ” Proof of Incompetence

Documented evidence of GPT-5's failures and regressions from previous models

⚠️ CRITICAL: These are documented, verified failures

Every example below has been reported by multiple sources and verified by tech publications. This isn't opinionβ€”it's evidence.

πŸ“Š Failure Statistics

15+

Documented Math Errors

50+

Geography Mistakes

100%

Chart Errors in Demo

24 hrs

Time to Crisis

🚨 CRITICAL
Elementary Math Failure

GPT-5 Can't Solve Basic Algebra

πŸ“… Documented Aug 8, 2025 πŸ” Multiple Sources ⚑ Immediate Failure

Multiple users and publications documented GPT-5's failure at elementary algebra that any middle school student could solve.

Problem: 5.9 = x + 5.11
GPT-5 Answer: -0.2
Correct Answer: 0.79
"It also failed on a simple algebra arithmetic problem that elementary schoolers could probably nail, 5.9 = x + 5.11." - VentureBeat

This represents a catastrophic regression from previous models that could handle basic arithmetic correctly. The error was documented by OpenAI's own developer community and reported across multiple tech publications.

OpenAI Developer Forum Report

Related Articles

🚨 CRITICAL
Number Comparison Fail

Claims 9.11 > 9.9

πŸ“… Documented Aug 8, 2025 πŸ” Social Media Evidence ⚑ Basic Logic Failure

Users posted screenshots showing GPT-5 making elementary mistakes in number comparison that demonstrate fundamental logical failures.

Which is larger: 9.11 or 9.9?
GPT-5 Answer: 9.11 is greater than 9.9
Correct Answer: 9.9 is greater than 9.11
"One user on X sarcastically posted 'AGI is here, everybody' alongside screenshots showing GPT-5 making elementary mistakes: claiming 9.11 is greater than 9.9" - Final Round AI

This type of error is particularly damaging because it shows GPT-5 failing at the most basic numerical reasoning that any calculator or computer system should handle flawlessly.

Full Analysis at Final Round AI
⚠️ MAJOR
Geography Catastrophe

Invents Fictional US States

πŸ“… Documented Aug 8, 2025 πŸ” Futurism Report ⚑ Creative Geography

When asked to create a simple US map, GPT-5 invented completely fictional state names that don't exist anywhere in reality.

Request: Draw a map of the US with state names
GPT-5 Created: "West Wigina," "Delsware," "Fiorata," "Rhoder land," "Tonnessee," "Mississipo"
Expected: Actual state names like West Virginia, Delaware, Florida, Rhode Island, Tennessee, Mississippi
"Ever think of visiting 'West Wigina,' 'Delsware,' 'Fiorata,' or 'Rhoder land'? Or maybe 'Tonnessee' and 'Mississipo?'" - Futurism

This represents a significant regression from GPT-4o's geographic knowledge and demonstrates how GPT-5's training may have actually corrupted basic factual information.

Read Full Report at Futurism
⚠️ MAJOR
State Labeling Fail

Labels Oklahoma as "Gelahbrin"

πŸ“… Documented Aug 8, 2025 πŸ” Multiple Publications ⚑ Map Generation Fail

Multiple users documented GPT-5 creating maps with completely nonsensical state labels, including renaming Oklahoma as the fictional "Gelahbrin."

Request: Create a map with Oklahoma labeled
GPT-5 Output: Labeled Oklahoma as "Gelahbrin"
Expected: Oklahoma
"Social media lit up with screenshots of GPT-5's bizarre errors, like maps labeling Oklahoma as 'Gelahbrin'" - Quartz

This error spread rapidly across social media as evidence of GPT-5's regression in basic geographic knowledge, something that should be fundamental for any AI system.

Source: Quartz Coverage
🚨 CRITICAL
Demo Catastrophe

Official Launch Demo Charts Were Completely Wrong

πŸ“… August 7, 2025 πŸ” Live Demonstration ⚑ Public Humiliation

OpenAI's own launch presentation contained catastrophically inaccurate charts that made GPT-5 look worse than it claimed to be.

Chart Issue: Coding Benchmark Comparison
GPT-5 (52.8%): Nearly twice as tall as o3 (69.1%)
Expected: 52.8% bar should be shorter than 69.1% bar
"The most egregious example is a bar graph comparing coding benchmark scores for GPT-5 compared to older models. Somehow, the bar for GPT-5's score of 52.8 percent accuracy is nearly twice as tall as the bar for a score of 69.1 percent" - Futurism

Sam Altman was forced to acknowledge the error on Twitter, writing "wow a mega chart screwup from us earlier." The fact that OpenAI couldn't even get their own presentation charts correct became a symbol of the entire botched launch.

Full Futurism Investigation
⚠️ MAJOR
Basic Counting Fail

Can't Count Words or Letters

πŸ“… Documented Aug 8, 2025 πŸ” User Testing ⚑ Elementary Task Failure

Users documented GPT-5 failing at basic counting tasks that any human could perform correctly.

Task: Count letters in "Tennessee"
GPT-5: Incorrect count
Correct: 9 letters (T-e-n-n-e-s-s-e-e)
Task: Identify 4th word in its own response
GPT-5: Said "the" was the 4th word
Correct: Should count accurately
"Screenshots showing GPT-5 making elementary mistakes... saying 'the' is the 4th word in its own response, and incorrectly counting letters in 'Tennessee'" - Final Round AI

These failures represent basic cognitive tasks that should be trivial for any AI system claiming to have "PhD-level intelligence."

Detailed Analysis
🚨 CRITICAL
Mathematical Proof Failure

Gets Mathematical Proof Wrong: 8.888... β‰  9

πŸ“… August 8, 2025 πŸ” Data Scientist Testing ⚑ Fundamental Math Error

Data scientist Colin Fraser documented GPT-5 failing a basic mathematical proof about repeating decimals.

Question: Does 8.888... (repeating) equal 9?
GPT-5: No, they are not equal
Correct: No, 8.888... β‰  9 (GPT-5 was actually right, but handled the proof poorly)
"Data scientist Colin Fraser posted screenshots showing GPT-5 getting a math proof wrong (whether 8.888 repeating is equal to 9 β€” it is of course, not)" - VentureBeat

While GPT-5 got the final answer correct, its reasoning and proof methodology were flawed, showing it doesn't understand the mathematical concepts it's working with.

VentureBeat Coverage
⚠️ MAJOR
Historical Fiction

Creates Fictional Presidents

πŸ“… Documented Aug 8, 2025 πŸ” Timeline Generation ⚑ Historical Hallucination

When asked to create historical timelines, GPT-5 invented completely fictional US presidents.

Request: Create timeline of US presidents
GPT-5 Output: Included "Willian H. Brusen" as a president
Expected: Only real historical US presidents
"Social media lit up with screenshots of GPT-5's bizarre errors... and timelines featuring fictional presidents like 'Willian H. Brusen.'" - Quartz

This demonstrates how GPT-5's training process may have corrupted basic historical knowledge, creating confident but completely false information.

Quartz Documentation
🚨 CRITICAL
Performance Regression