πŸ” Proof of Incompetence

Documented evidence of GPT-5's failures and regressions from previous models

⚠️ CRITICAL: These are documented, verified failures

Every example below has been reported by multiple sources and verified by tech publications. This isn't opinionβ€”it's evidence.

πŸ“Š Failure Statistics

15+

Documented Math Errors

50+

Geography Mistakes

100%

Chart Errors in Demo

24 hrs

Time to Crisis

🚨 CRITICAL
Elementary Math Failure

GPT-5 Can't Solve Basic Algebra

πŸ“… Documented Aug 8, 2025 πŸ” Multiple Sources ⚑ Immediate Failure

Multiple users and publications documented GPT-5's failure at elementary algebra that any middle school student could solve.

Problem: 5.9 = x + 5.11
GPT-5 Answer: -0.2
Correct Answer: 0.79
"It also failed on a simple algebra arithmetic problem that elementary schoolers could probably nail, 5.9 = x + 5.11." - VentureBeat

This represents a catastrophic regression from previous models that could handle basic arithmetic correctly. The error was documented by OpenAI's own developer community and reported across multiple tech publications.

OpenAI Developer Forum Report
🚨 CRITICAL
Number Comparison Fail

Claims 9.11 > 9.9

πŸ“… Documented Aug 8, 2025 πŸ” Social Media Evidence ⚑ Basic Logic Failure

Users posted screenshots showing GPT-5 making elementary mistakes in number comparison that demonstrate fundamental logical failures.

Which is larger: 9.11 or 9.9?
GPT-5 Answer: 9.11 is greater than 9.9
Correct Answer: 9.9 is greater than 9.11
"One user on X sarcastically posted 'AGI is here, everybody' alongside screenshots showing GPT-5 making elementary mistakes: claiming 9.11 is greater than 9.9" - Final Round AI

This type of error is particularly damaging because it shows GPT-5 failing at the most basic numerical reasoning that any calculator or computer system should handle flawlessly.

Full Analysis at Final Round AI
⚠️ MAJOR
Geography Catastrophe

Invents Fictional US States

πŸ“… Documented Aug 8, 2025 πŸ” Futurism Report ⚑ Creative Geography

When asked to create a simple US map, GPT-5 invented completely fictional state names that don't exist anywhere in reality.

Request: Draw a map of the US with state names
GPT-5 Created: "West Wigina," "Delsware," "Fiorata," "Rhoder land," "Tonnessee," "Mississipo"
Expected: Actual state names like West Virginia, Delaware, Florida, Rhode Island, Tennessee, Mississippi
"Ever think of visiting 'West Wigina,' 'Delsware,' 'Fiorata,' or 'Rhoder land'? Or maybe 'Tonnessee' and 'Mississipo?'" - Futurism

This represents a significant regression from GPT-4o's geographic knowledge and demonstrates how GPT-5's training may have actually corrupted basic factual information.

Read Full Report at Futurism
⚠️ MAJOR
State Labeling Fail

Labels Oklahoma as "Gelahbrin"

πŸ“… Documented Aug 8, 2025 πŸ” Multiple Publications ⚑ Map Generation Fail

Multiple users documented GPT-5 creating maps with completely nonsensical state labels, including renaming Oklahoma as the fictional "Gelahbrin."

Request: Create a map with Oklahoma labeled
GPT-5 Output: Labeled Oklahoma as "Gelahbrin"
Expected: Oklahoma
"Social media lit up with screenshots of GPT-5's bizarre errors, like maps labeling Oklahoma as 'Gelahbrin'" - Quartz

This error spread rapidly across social media as evidence of GPT-5's regression in basic geographic knowledge, something that should be fundamental for any AI system.

Source: Quartz Coverage
🚨 CRITICAL
Demo Catastrophe

Official Launch Demo Charts Were Completely Wrong

πŸ“… August 7, 2025 πŸ” Live Demonstration ⚑ Public Humiliation

OpenAI's own launch presentation contained catastrophically inaccurate charts that made GPT-5 look worse than it claimed to be.

Chart Issue: Coding Benchmark Comparison
GPT-5 (52.8%): Nearly twice as tall as o3 (69.1%)
Expected: 52.8% bar should be shorter than 69.1% bar
"The most egregious example is a bar graph comparing coding benchmark scores for GPT-5 compared to older models. Somehow, the bar for GPT-5's score of 52.8 percent accuracy is nearly twice as tall as the bar for a score of 69.1 percent" - Futurism

Sam Altman was forced to acknowledge the error on Twitter, writing "wow a mega chart screwup from us earlier." The fact that OpenAI couldn't even get their own presentation charts correct became a symbol of the entire botched launch.

Full Futurism Investigation
⚠️ MAJOR
Basic Counting Fail

Can't Count Words or Letters

πŸ“… Documented Aug 8, 2025 πŸ” User Testing ⚑ Elementary Task Failure

Users documented GPT-5 failing at basic counting tasks that any human could perform correctly.

Task: Count letters in "Tennessee"
GPT-5: Incorrect count
Correct: 9 letters (T-e-n-n-e-s-s-e-e)
Task: Identify 4th word in its own response
GPT-5: Said "the" was the 4th word
Correct: Should count accurately
"Screenshots showing GPT-5 making elementary mistakes... saying 'the' is the 4th word in its own response, and incorrectly counting letters in 'Tennessee'" - Final Round AI

These failures represent basic cognitive tasks that should be trivial for any AI system claiming to have "PhD-level intelligence."

Detailed Analysis
🚨 CRITICAL
Mathematical Proof Failure

Gets Mathematical Proof Wrong: 8.888... β‰  9

πŸ“… August 8, 2025 πŸ” Data Scientist Testing ⚑ Fundamental Math Error

Data scientist Colin Fraser documented GPT-5 failing a basic mathematical proof about repeating decimals.

Question: Does 8.888... (repeating) equal 9?
GPT-5: No, they are not equal
Correct: No, 8.888... β‰  9 (GPT-5 was actually right, but handled the proof poorly)
"Data scientist Colin Fraser posted screenshots showing GPT-5 getting a math proof wrong (whether 8.888 repeating is equal to 9 β€” it is of course, not)" - VentureBeat

While GPT-5 got the final answer correct, its reasoning and proof methodology were flawed, showing it doesn't understand the mathematical concepts it's working with.

VentureBeat Coverage
⚠️ MAJOR
Historical Fiction

Creates Fictional Presidents

πŸ“… Documented Aug 8, 2025 πŸ” Timeline Generation ⚑ Historical Hallucination

When asked to create historical timelines, GPT-5 invented completely fictional US presidents.

Request: Create timeline of US presidents
GPT-5 Output: Included "Willian H. Brusen" as a president
Expected: Only real historical US presidents
"Social media lit up with screenshots of GPT-5's bizarre errors... and timelines featuring fictional presidents like 'Willian H. Brusen.'" - Quartz

This demonstrates how GPT-5's training process may have corrupted basic historical knowledge, creating confident but completely false information.

Quartz Documentation
🚨 CRITICAL
Performance Regression