Anthropic Identifies Three Product-Layer Changes Behind Claude Code Quality Decline, Not Model Issue

Gate News message, April 23 — Anthropic’s engineering team confirmed that the Claude Code quality degradation reported by users over the past month stemmed from three independent product-layer changes, not from API or underlying model issues. The three problems were fixed on April 7, April 10, and April 20 respectively, with the final version now at v2.1.116.

The first change occurred on March 4, when the team reduced the default reasoning effort level for Claude Code from “high” to “medium” to address occasional extreme latency spikes in Opus 4.6 under high reasoning intensity. After widespread user complaints about reduced performance, the team reverted the change on April 7. The current default is now “xhigh” for Opus 4.7 and “high” for other models.

The second issue was a bug introduced on March 26. The system was designed to clear old reasoning records after conversation inactivity exceeded one hour to reduce session recovery costs. However, a flaw in implementation caused the clearing to execute repeatedly on every subsequent turn rather than once, causing the model to progressively lose prior reasoning context. This manifested as increasing forgetfulness, repeated operations, and abnormal tool invocations. The bug also resulted in cache misses on every request, accelerating user quota consumption. Two unrelated internal experiments masked the reproduction conditions, extending the debugging process to over a week. After fixing on April 10, the team reviewed problematic code using Opus 4.7 and found that Opus 4.7 could identify the bug while Opus 4.6 could not.

The third change launched on April 16 alongside Opus 4.7. The team added instructions to the system prompt to reduce redundant output. Internal testing over several weeks showed no regression, but post-launch interaction with other prompts degraded coding quality. Extended evaluation revealed a 3% performance drop in both Opus 4.6 and 4.7, leading to a rollback on April 20.

These three changes affected different user groups at different times, and their combined effect created widespread and inconsistent quality decline, complicating diagnosis. Anthropic stated it will now require more internal employees to use the same public build version as users, run full model evaluation suites for every system prompt modification, and implement staged rollout periods. As compensation, Anthropic has reset usage quotas for all subscription users.

免责声明:本页面信息可能来自第三方,不代表 Gate 的观点或意见。页面显示的内容仅供参考,不构成任何财务、投资或法律建议。Gate 对信息的准确性、完整性不作保证,对因使用本信息而产生的任何损失不承担责任。虚拟资产投资属高风险行为,价格波动剧烈,您可能损失全部投资本金。请充分了解相关风险,并根据自身财务状况和风险承受能力谨慎决策。具体内容详见声明

相关文章

Alphabet 财报超预期,GOOG 上涨 6% 创新高

Alphabet 第一季营收与EPS皆超预期,云端营收200亿美元,未交货订单倍增至4,600亿美元。资本支出上调至1,850亿美元,2027年将高于2026年。Gemini Enterprise用户增长、搜索查询创新高,AI解答影响广告模式。股价盘后上涨约6%至370美元,创历史新高。

鏈新聞abmedia24 分钟前

英伟达深度学习副总裁认为 AI 运算支出超越人力薪资成本

英伟达深度学习副总裁表示,维持AI模型运算的成本远高于人力薪资,显示AI在实务上未必能降低劳动成本。MIT 2024研究指出,AI自动化在约23%岗位具备经济效益,其余77%仍以人力为主。全球科技巨头在AI基础建设的投资居高不下,短期财务压力与裁员潮并存;但若通过更大规模的稳定化与较低监督成本,长期仍有降本与经济效益的可能。

鏈新聞abmedia27 分钟前

AI 金融平台 Rogo 在不到 3 个月内完成 D 轮融资,由 Kleiner Perkins 牵头

据《Beating》报道,面向高频金融场景的人工智能平台 Rogo 已于 2026 年 4 月完成一轮 $160 百万美元的 D 轮融资,由 Kleiner Perkins 牵头,Sequoia、Thrive Capital、Khosla Ventures 以及 J.P. Morgan 参与。该轮融资距离公司在 1 月下旬完成一轮 百万美元的 C 轮融资不足三个月,使总融资额超过 百万。 Rogo 同步推出 Felix,这是一款面向金融从业者的人工智能代理平台。该平台使投资银行家和分析师能够将包括财务建模、报告分析以及演示文稿制作在内的多步骤任务委托给 Felix,从而释放时间用于高杠杆的客户工作。目前,该平台已被超过 35,000 名专业人士在 250 多家顶级投资银行、资产管理公司和私募股权公司中使用。

GateNews7小时前

中国以技术与数据安全担忧为由,阻止 Meta 支持的 Manus AI 收购

据 PANews 报道,4 月 29 日,中国国家发展和改革委员会投资安全审查办公室禁止对 Manus 项目的外国收购,并要求终止该交易。Manus 号称是世界首个通用人工智能代理,曾

GateNews7小时前

阿里云将 DeepSeek-V4-Pro 隐式缓存定价下调至 4 月 29 日每百万 tokens 1 元

据阿里云介绍,其百炼平台将下调 DeepSeek-V4-Pro 模型隐式缓存 (Implicit Cache) 的定价:自北京时间 2026 年 4 月 29 日 23:59:59 起,每百万 tokens 价格为 1 元。只有当请求命中缓存时才适用隐式缓存;缓存的输入 tokens 按 cached_token 费率计费,而未命中缓存的输入 tokens 则按标准 input_token 费率计费。此次调整仅影响隐式缓存的定价;基础模型推理费率保持不变。

GateNews8小时前
评论
0/400
暂无评论