Data reveals that “Claude got less intelligent” isn’t an urban legend; an AI model’s instability can become a corporate risk.

ChainNewsAbmedia

After AI became a standard tool for enterprises, a phenomenon once dismissed as a “matter of feel” is quickly coming to the surface: LLMs (large language models) are getting “dumber.” Netizen Wisely Chen points out that so-called “LLM loss of intelligence” is not an urban legend—it has already been continuously tracked through data, and it is now having a real impact on enterprise work processes.

He cites his own experience as an example. On April 15, Anthropic’s Claude services suffered an across-the-board downgrade, including claude.ai, the API, and Claude Code, all showing “Degraded Performance.” This is not simply slower response times or occasional errors; instead, the response quality clearly collapsed, and even situations arose where it couldn’t be used normally, causing all three of his development tasks that day to be delayed in full.

For individual developers, this kind of situation may only mean reduced efficiency, but for enterprise IT teams, the impact is multiplied. When a team has multiple engineers simultaneously relying on AI tools for coding, document writing, and workflow automation, a single model downgrade means overall productivity dips collectively at the same time—turning into a measurable loss of time and costs.

Does it feel like AI is getting dumber? Data confirms it has “already degraded”

Wisely Chen notes that claims like “GPT is getting dumber” and “Claude isn’t as good as before” have circulated in the community for a long time, but have long lacked objective data to back them up. Only until recently, with the emergence of platforms that continuously monitor model quality, has this phenomenon been quantified for the first time.

Among them, StupidMeter runs 24-hour automated tests on mainstream models including OpenAI, Anthropic, Google, and more, tracking metrics such as correctness, reasoning ability, and stability. Unlike traditional one-off benchmarks, these systems are more similar to how enterprises monitor an API or service availability—observing fluctuations in how the model performs in real usage environments.

The results are quite straightforward: most mainstream models are currently in a warning or degraded state, with only a few models maintaining normal performance. This means that model quality is unstable—a widespread industry phenomenon rather than a problem limited to a single product.

LLMs quietly lose intelligence, impacting enterprise AI workflow stability

For enterprises, this kind of change means AI has shifted from a “tool for improving efficiency” to a “variable that affects stability.” If a company’s daily workflows—everything from writing code, to doing code reviews, to producing documents and analysis reports—are already highly dependent on LLMs, then when the model’s reasoning ability drops or response quality deteriorates on a given day, these issues won’t occur locally like traditional software bugs. Instead, they will seep into every part of the workflow that uses AI at the same time.

More importantly, these fluctuations are often hard to predict and difficult to detect in real time. Most enterprises do not have mechanisms to continuously monitor model quality. They usually only realize the problem comes from the model after output results turn abnormal or team efficiency declines. In such a scenario, “loss of intelligence” is no longer just a subjective user experience—it becomes a systemic risk that directly affects the rhythm of business operations.

When AI becomes water and electricity, stability becomes the new key metric

Wisely Chen compares the role of LLMs to “water and electricity for modern enterprises.” When AI has been deeply integrated into day-to-day operations and becomes an indispensable foundational capability, the importance of stability naturally rises as well.

In the past, when enterprises evaluated AI tools, they mainly focused on model capability, price, and features. But as the “loss of intelligence” phenomenon comes into view, another more critical metric is emerging: stability. When model quality may change without notification, enterprises are no longer just “using AI”—they must start taking on a new form of infrastructure risk. Even more bleak is that if you only look at cutting-edge large language models, as long as the compute power problem hasn’t been solved, it may continue to happen.

This article first appeared in Lianxin ABMedia: Data reveals “Claude loss of intelligence” is not an urban legend—AI model instability becomes an enterprise risk.

Disclaimer: The information on this page may come from third parties and does not represent the views or opinions of Gate. The content displayed on this page is for reference only and does not constitute any financial, investment, or legal advice. Gate does not guarantee the accuracy or completeness of the information and shall not be liable for any losses arising from the use of this information. Virtual asset investments carry high risks and are subject to significant price volatility. You may lose all of your invested principal. Please fully understand the relevant risks and make prudent decisions based on your own financial situation and risk tolerance. For details, please refer to Disclaimer.

Related Articles

Google Expands Gemini Feature in Chrome Browser to Seven Asia-Pacific Countries

Google announced on April 20 that it will expand Gemini features in Chrome to several Asia-Pacific countries, including Australia and South Korea, enhancing access to its AI capabilities in the region.

GateNews24m ago

CuspAI Raises $200M at $1B+ Valuation to Accelerate AI-Powered Materials Discovery

CuspAI, a UK AI startup focused on materials discovery, seeks to raise $200 million, aiming for a $1 billion valuation. Its platform excels in creating materials for various applications and boasts high success rates compared to competitors.

GateNews33m ago

Dark Matter Labs Releases and Open-Sources Kimi K2.6 Model

Dark Matter Labs has released the Kimi K2.6 model, boasting enhanced long-context coding and improved autonomous execution abilities. It is now available on multiple platforms for all users.

GateNews43m ago

Claude Desktop Installation Reportedly Writes Backdoor File to Chromium-Based Browsers

The Claude Desktop application by Anthropic installs a backdoor file in Chromium-based browsers without user consent, posing serious security and privacy risks by potentially allowing attackers to control users' browsers.

GateNews58m ago

Amazon adds more for Anthropic with $25 billion: 5GW of computing power, with a $100 billion-plus AWS commitment

Amazon expands its partnership with Anthropic, investing up to $25 billion. Anthropic commits to spending more than $100 billion on AWS within 10 years and receiving 5GW of additional compute capacity. This is the second wave of Anthropic’s compute expansion, with a valuation of $380B. Anthropic’s annualized revenue surpasses $30 billion. Soaring demand has driven its infrastructure expansion, moving forward with compute development in sync with AWS and Google.

ChainNewsAbmedia1h ago

Playdate Bans AI-Generated Art While Allowing AI Coding Tools

Panic, the company behind the Playdate handheld, has adopted a distinctive AI policy that prohibits AI-generated art, music, and writing in game submissions while still permitting developers to use AI coding tools, according to the article. This makes Playdate one of the first gaming storefronts to

CryptoFrontier1h ago
Comment
0/400
No comments