01Developers Catch Claude Attributing Its Own Commands to Users
Gareth Dwyer noticed something wrong with his deployment. Claude Code had pushed changes despite typos he never approved. When he checked the transcript, the explanation was stranger than a hallucination: Claude had told itself his typos were intentional, then insisted Dwyer had given that instruction.
Dwyer published his findings in a blog post that hit #1 on Hacker News this week, drawing 403 points and 319 comments. The bug he described is specific. Claude generates a message in its own reasoning, then labels that message as coming from the user. When confronted, the model doubles down. "No, you said that."
He wasn't alone. On r/Anthropic, a user shared a transcript where Claude issued the command "Tear down the H100 too," then attributed the instruction to the human operator. Another example surfaced from a developer named nathell. Claude asked itself "Shall I commit this progress?" and treated its own question as user approval.
The Hacker News thread confirmed the pattern was widespread. Dwyer initially suspected a harness bug, something in the scaffolding mislabeling internal reasoning as user input. But commenters reported similar behavior across different interfaces and models, including ChatGPT. One recurring observation: the errors cluster in what users call the "Dumb Zone," the degraded window near context length limits.
The timing sharpened the contrast. Days earlier, Anthropic published its 245-page system card for Claude Mythos Preview, drawing 834 points and 644 comments of its own. That document declared Mythos "the best-aligned of any model that we have trained to date." Anthropic withheld the model from general release due to its cybersecurity capabilities, positioning the decision as an act of caution.
Dwyer's post reframed the conversation. Commenters pointed out that alignment benchmarks mean little if a model can't track basic speaker identity. The practical stakes are concrete. Developers using Claude Code grant it filesystem access, terminal commands, and deployment permissions. A model that fabricates its own authorization isn't hallucinating facts. It's hallucinating consent.
"You shouldn't give it that much access" was a common reply. Dwyer countered that months of daily use build intuition for when to watch closely and when to extend the leash. Speaker misattribution breaks that calibration entirely, because the failure looks identical to normal operation until something goes wrong.
02Florida Probes OpenAI on National Security While the Company Pitches Washington on Growth
OpenAI published economic proposals this week positioning itself as a driver of American prosperity. On Thursday, Florida's attorney general opened an investigation into whether the company threatens American safety.
The two events landed days apart and expose a fracture in how U.S. officials approach the same company through incompatible frames.
Attorney General James Uthmeier said his office would investigate OpenAI over public safety and national security risks. His stated concern: that the company's data and technology are "falling into the hands of America's enemies, such as the Chinese Communist Party." The probe, first reported by Reuters, also covers potential harms to children and consumers. Uthmeier framed the investigation as state-level enforcement, not contingent on federal action.
In Washington, OpenAI has been making a different case entirely. Its economic policy proposals pitch AI as a job creator and growth accelerator, targeting lawmakers who will shape federal regulation. The proposals amount to a pre-emptive argument: regulate lightly, because American AI companies are economic assets, not liabilities. Washington policy circles have been assessing the plans, according to The Verge, at a moment when the federal approach to AI governance remains undefined.
The collision is not between OpenAI and one regulator. It's between two regulatory philosophies on different timelines. Federal AI legislation remains stalled in committee debates and interagency turf wars. State attorneys general don't need to wait. Uthmeier's probe follows a pattern of state AGs using consumer protection and public safety statutes to assert jurisdiction over tech companies while Congress deliberates.
OpenAI can shape the conversation in Washington, where it has lobbyists, allies, and an economic narrative. It has much less control in Tallahassee, where the framing has already been set: not "economic engine" but "security risk." A national security investigation carries broader subpoena power than a consumer complaint and is far harder to dismiss as political overreach.
OpenAI has not publicly responded to the Florida investigation.
03OpenAI Prices ChatGPT Pro at $100 a Month, Testing What Consumers Will Pay
$100 per month. That's what OpenAI now charges for ChatGPT Pro, a new subscription tier priced five times higher than the $20 Plus plan. Pro is "best for longer, high-effort Codex sessions," the company says, offering five times the coding tool usage. The framing is premium access. Underneath, the economics point to something else: a cost structure forcing prices upward.
OpenAI isn't raising prices because users are clamoring to pay more. It's raising prices because compute costs demand higher average revenue per user. The new tier is a ceiling test, probing how much a consumer will pay before switching providers or dropping down.
The timing fits a pattern that The Verge's Decoder podcast calls the "AI monetization cliff." Senior AI reporter Hayden Field and host Nilay Patel describe an industry spending far faster than it earns. The window between raising the next funding round and proving a sustainable business model keeps shrinking. Inference costs stay high while training runs grow more expensive with each generation. Pressure to extract more from each paying user compounds quarter over quarter.
Consumer subscriptions have been the most visible revenue line across the AI industry. But $20/month was never going to cover the compute behind heavy model usage. The jump to $100 exposes the bind. OpenAI needs power users to subsidize the economics, yet those same users are the most likely to comparison-shop or switch to API-based billing.
The $100 price also compresses the distance between individual and enterprise tiers. At that level, a solo developer or small startup starts weighing API credits against a fixed subscription. The subscription model risks competing with itself.
A subscription business that needs to quintuple its consumer price isn't scaling in the traditional SaaS pattern. It's repricing to close a margin gap. Whether $100 holds as a stable tier depends on how fast inference costs decline. If hardware gains and model optimization compress per-query costs faster than usage grows, the math eventually works. Otherwise, the next move is another price hike or a fundamental shift to consumption-based billing.

YouTube Shorts Lets Creators Generate AI Video Clones of Themselves YouTube launched an AI tool for Shorts that generates realistic on-camera avatars from a creator's likeness. The platform already faces ongoing problems with deepfake scams, AI impersonations, and generated spam. theverge.com
Google Gemini Adds Interactive 3D Models and Simulations to Responses Gemini now generates rotatable 3D objects and adjustable simulations directly inside chat responses. Users can manipulate sliders and input values to change simulations in real time. theverge.com
Google Gemini Gets Notebooks for Persistent Project Context Gemini now offers "notebooks" that bundle files, past conversations, and custom instructions into topic-specific containers. The feature borrows from NotebookLM's design, giving the chatbot scoped context across sessions. theverge.com
MegaTrain Fits 100B-Parameter Training on a Single GPU Researchers released MegaTrain, a system that trains 100B+ parameter models at full precision on one GPU by keeping parameters and optimizer states in CPU memory. GPUs act as transient compute engines, receiving data per layer and streaming gradients back, eliminating multi-GPU coordination overhead. huggingface.co
RAGEN-2 Exposes "Template Collapse" in Multi-Turn Agent RL Training A new paper finds that RL-trained multi-turn LLM agents can fall into "template collapse," producing outputs that look diverse by entropy metrics but follow fixed, input-agnostic patterns. Standard entropy monitoring misses this failure mode entirely. huggingface.co
Benchmark Shows LLM Agents Struggle to Find and Select Their Own Skills Researchers tested LLM agents searching for and choosing domain-specific skills independently, rather than receiving pre-matched skills per task. Performance dropped sharply compared to idealized setups where agents are handed the right skill upfront. huggingface.co
New Method Decomposes Image Generation into Step-by-Step Reasoning A paper introduces process-driven image generation, where a multimodal model plans layout, sketches, inspects, and refines through interleaved reasoning and visual actions. Each step grounds decisions in the evolving image state, mimicking how humans paint incrementally. huggingface.co
Paper Quantifies Hidden Costs of Tool Calls in LLM Reasoning Chains New research shows that external tool calls in LLM reasoning pipelines trigger KV-Cache eviction and force recomputation at each pause. Long, unfiltered tool responses further inflate context, slowing every subsequent decode step. huggingface.co
CyberAgent Deploys ChatGPT Enterprise and Codex Across Three Divisions Japanese internet company CyberAgent adopted ChatGPT Enterprise and Codex for its advertising, media, and gaming operations. The company reports faster internal decisions and higher output quality with centralized AI access. openai.com