Skip to content

HackerNews AI β€” 2026-04-17

1. What People Are Talking About

1.1 Opus 4.7 Token Cost Shock and the Anthropic Trust Deficit πŸ‘•

The day's dominant story β€” 480 points, 320 comments β€” was a detailed measurement of Claude Opus 4.7's tokenizer costs that revealed Anthropic's published range understates the real-world impact. aray07 shared a Claude Code Camp analysis showing that on real Claude Code content (CLAUDE.md files, prompts, diffs), the new tokenizer uses 1.325x more tokens on a weighted basis, with technical docs hitting 1.47x β€” well above Anthropic's stated "1.0 to 1.35x" range (post). The article used Anthropic's own count_tokens endpoint across 19 samples spanning real-world and synthetic content. Code is hit hardest: chars-per-token on TypeScript dropped from 3.66 to 2.69. CJK content barely changed (~1.01x). Same sticker price, same quota, but your context window burns through faster.

This landed alongside three other Opus 4.7-related posts that collectively painted a picture of a degrading developer experience. birdculture shared a blog post documenting a pattern of adversarial changes: removal of clear-context-on-plan-accept, cache TTL quietly cut from 1 hour to 5 minutes, 3rd-party programs blocked from Pro/Max plan tokens, and extended thinking budgets removed entirely in Opus 4.7 (post). The post argues these are capacity management moves that "impact the quality of life for all users of Claude Code β€” even when you are paying for tokens through the API." Anthropic itself published official guidance recommending a new xhigh effort level and fewer interactive turns β€” essentially asking users to adapt their workflow to the model's new economics (post).

Discussion insight: tabbott pushed back on the cost focus, arguing that at current price points "human time spent redirecting AI coding agents... remains dramatically more expensive than the token cost" and that "$200/month is an expensive hobby, but it's negligible as a business expense." pdp drew an analogy to display resolution: "comparing an 8K display to a 16K display... the difference is imperceptible." namnnumbr warned the analysis assumes Opus 4.7 uses the same reasoning trajectory as 4.6: "I assume it not to be the case, given that Opus 4.7 on Low thinking is strictly better than Opus 4.6 on Medium." speedgoose noted the GitHub Copilot multiplier went from 3 to 7.5, calling it "Microsoft wanting to lose money slightly slower."

Prior-day comparison: On April 15, the Claude reliability crisis manifested as outages and 500 errors; today it has shifted from availability to economics. The frustration has evolved from "I can't use it" to "I'm paying more for less."

1.2 AI Agent Readiness β€” and the Backlash Against It πŸ‘•

WesSouza shared isitagentready.com, a Cloudflare-affiliated tool that scans websites across five categories β€” discoverability, content accessibility, bot access control, protocol discovery (MCP, WebMCP, OAuth), and agentic commerce (x402, UCP, ACP) (post). With 92 points and 159 comments, it generated the day's second-largest discussion β€” overwhelmingly skeptical.

The top comment, by Mordisquitos, captured the sentiment: "AI industry: 'AI agents will soon be able to do any white-collar human job!' Also AI industry: 'Please make sure your website is adapted so that AI agents are able to use it.'" pickleglitch wanted the opposite tool β€” one that shows "how well my site is protected from being accessed by AI agents." Multiple users reported their WAF blocking the scanner and treated the 403 as a badge of honor. xnx dismissed the concept as "SEO hucksters trying to pivot their obsolete services into 'GEO'."

Despite the backlash, Brajeshwar shared a New Stack interview with AWS's Clare Liguori, an MCP core maintainer, about Amazon's expanding MCP investment β€” contributing Tasks and Elicitations to the spec and using AWS managed MCP servers as an "experimental playground" for draft features (post). The article noted Amazon expanded its Kiro AI dev tool to all roles after demand exceeded engineering-only expectations.

Prior-day comparison: April 15 saw MCP discussed as a protocol with debugging problems. Today the discourse has split: enterprise adoption accelerating (Amazon, Cloudflare) while individual developers resist making their sites agent-accessible.

1.3 Hardware Meets AI: Domain-Specific Agent Failures πŸ‘•

fizz_buzz built MCP servers for a LeCroy oscilloscope and SPICE simulator, enabling Claude Code to close the loop between simulation and real hardware measurement (post). With 116 points and 31 comments, this Show HN showcased an ambitious workflow β€” but the comments were more instructive than the demo.

iterateoften reported that Claude "completely hallucinated the capabilities of the board and made some pretty crazy claims like I had just stumbled onto the secret hardware billion dollar project that every home needed. None of the boards worked." Eextra953 confirmed agents "fall apart when trying to do anything other than design the simplest circuit" because "they have no concept of the physics behind a circuit." The practical solution, shared by andrewklofas, was to stop letting Claude read domain files directly and instead run Python analyzers that output JSON β€” "Claude just reads the JSON, problem mostly went away."

The pattern: domains with physical constraints (circuits, mechanical systems) require structured intermediaries between agents and domain tools. Raw file access invites hallucination.

1.4 AI Slop Flooding Open Source πŸ‘’

motakuk shared an Archestra.ai blog post detailing how AI bots poisoned their open-source repo after they posted a $900 bounty (post). The post described an issue ballooning to 253 comments from AI accounts, 27 untested PRs for a single feature, and a team member spending "half a day every week cleaning AI garbage." Their solution β€” blocking all non-onboarded contributors β€” is self-described as a "nuclear option" that could alienate legitimate newcomers, but they concluded: "we value quality over quantity. We don't value metrics pumped by AI slop."

Prior-day comparison: April 15 discussed AI-driven vulnerability discovery forcing open-source to close. Today's variant is AI-driven contribution spam forcing open-source to gate access. Both erode the open contribution model, but from different vectors.

1.5 The Solo Business Dream Meets AI Reality πŸ‘’

fnoef asked whether building a solo business is possible, noting the "push of vibe coding across the board" as added motivation to escape salaried work (post). The 50-comment thread was unusually substantive. dx-800 shared a 7-year journey from a Classic ASP intranet app to a SaaS serving 80 mobile home dealerships in 13 states: "The programming is the fun part... Selling is the hard part." 0xmattf documented multiple failures β€” Shopify stores, martial arts SaaS, chess browser extension β€” concluding: "Abandon the idea of making money, and it becomes more enjoyable." adzicg recommended Bill Aulet's "beachhead market" framework, emphasizing that "word of mouth and happy customers will be your first best marketing strategy."


2. What Frustrates People

Opus 4.7 Token Inflation and Silent Degradation

The day's highest-signal frustration. Real-world measurements show Claude Opus 4.7's tokenizer inflates token counts 1.21-1.47x across code and technical content, while Anthropic's published range (1.0-1.35x) understates the impact on typical Claude Code workflows (post). Combined with the removal of extended thinking budgets, cache TTL cuts, and third-party token access blocking (post), developers perceive a pattern of quiet capacity management at the expense of user experience. chmod775 provided a concrete code example where Opus 4.7 produced a 9-line overcomplicated eviction loop that could have been 5 lines, concluding: "I'm back to 4.6 for now" (post). Severity: High. Affects cost, quota burn rate, and code quality simultaneously.

Opus 4.7 Writing Quality Regression

limalabs found Opus 4.7 writing to be "sloppy, unprecise, very empty sentences" compared to 4.6, discovered mid-Master's thesis (post). muzani noted Anthropic doesn't benchmark writing quality despite a "very large and vocal user base using it for writing." The gap between coding optimization and writing quality creates a split: developers upgrading for code may find their non-code workflows degraded. Severity: Medium. Workaround exists (downgrade to 4.6 via web) but fragments tool usage.

AI Bots Poisoning Open Source Contributions

Archestra.ai's experience β€” 253 bot comments on a single issue, 27 untested PRs, half a day per week of cleanup β€” represents a growing operational cost for open-source maintainers (post). Their "nuclear option" of requiring onboarding before any interaction is a real barrier to legitimate new contributors. GitHub's contribution gating mechanisms were not designed for this failure mode. Severity: Medium. Disproportionately affects projects with bounties or high visibility.

Claude Code Usage Policy False Positives

sminchev hit a new Opus 4.7 content restriction when processing a personal email file: "Claude Code is unable to respond to this request, which appears to violate our Usage Policy" β€” triggered by real email addresses in an .eml file (post). Copying the content without addresses works fine. New safety gates that block legitimate local file processing undermine trust in the tool for professional workflows. Severity: Medium.

AI Data Training on Defunct Company Communications

AI companies purchasing Slack data from failed startups for training data raised ethical concerns (post). The Forbes article (by Anna Tong, 2026-04-16) documents the practice. kittikitti summarized the quality concern: "Slop in, slop out." Severity: Low (for developers), but signals a broader data-sourcing problem.


3. What People Wish Existed

Transparent Token Cost Accounting for Model Upgrades

The tokenizer measurement post (post) revealed that Anthropic's published range understates real-world costs. Developers want side-by-side cost comparisons before upgrading: same task, both tokenizers, showing total tokens and effective price change. namnnumbr specifically called for "Artificial Analysis' Intelligence Index" or "some other independent per-task cost analysis" rather than raw token counts. Opportunity: direct.

Agent-Proof Open Source Contribution Gating

Archestra.ai's half-day-per-week cleanup burden and "nuclear option" of contributor onboarding (post) points to a gap in GitHub's tooling. Maintainers need lightweight bot detection that doesn't block legitimate first-time contributors β€” something between "open to everyone" and "commit to main first." The reputation bot approach (London-Cat) and AI sheriff both had false positives. Opportunity: competitive.

Website Anti-Agent Readiness Scanner

The overwhelmingly negative reaction to isitagentready.com (post) produced a clear wish from pickleglitch: a tool that shows "how well my site is protected from being accessed by AI agents" with advice on locking it down further. Multiple users treating WAF 403 responses as success validates demand for the inverse of Cloudflare's tool. Opportunity: direct.

Structured Intermediaries for Hardware Domain Agents

The SPICE/oscilloscope discussion (post) converged on a pattern: agents cannot reliably read domain-specific file formats (KiCad schematics, SPICE netlists) directly, but work well when Python analyzers produce JSON summaries. andrewklofas built this for KiCad; the SPICE demo author built it for oscilloscopes. A generalizable framework for domain-file-to-JSON adapters, optimized for agent consumption, would serve the growing hardware-meets-AI community. Opportunity: aspirational.

Research-Augmented Coding Agents

Paper Lantern (post) demonstrated 30-80% improvement on 5 of 9 tasks by surfacing research techniques to coding agents. vunderba built a similar system independently. The wish is for coding agents that routinely consult recent research rather than relying solely on training data and web search. Opportunity: competitive.


4. Tools and Methods in Use

Tool Category Sentiment Strengths Limitations
Claude Code Coding Agent (+/-) Deep agentic reasoning, wide ecosystem, 1M context Opus 4.7 token inflation, removed extended thinking budgets, cache TTL cuts
Claude Opus 4.7 LLM (+/-) Better instruction following at lower effort levels, new xhigh setting 1.2-1.47x token inflation, writing quality regression, overcomplicated code output
Claude Opus 4.6 LLM (+) Stable, good writing quality, reliable code output Being superseded by 4.7 as default
MCP Agent Protocol (+/-) Cross-client, Amazon investing (Tasks, Elicitations), Cloudflare adopting Design flaw affecting 200K servers per researchers, debugging still painful
isitagentready.com Agent Readiness Scanner (-) Comprehensive 5-category check, Cloudflare-backed Deeply unpopular with HN audience; blocked by many WAFs
Codg Coding Agent Harness (+) Multi-model, async concurrency, TUI+web+desktop, local models Early stage, Go binary
SPICE / LeCroy MCP Hardware Integration (+/-) Closes simulation-to-hardware loop Agents hallucinate hardware capabilities without structured intermediaries
Paper Lantern Research MCP (+) 30-80% improvement on coding tasks, 2M+ papers Early, needs validation beyond benchmarks
Egregore Team Coordination (+) Git-backed memory, Claude Code hooks, /handoff /invite New, no external validation yet

The Claude Code ecosystem dominates today's discussion (24 mentions of "claude code" in the review set). The day's meta-story is the tension between Claude Code's expanding adoption and Anthropic's capacity management moves. Multiple independent tools address Claude Code quota tracking: micaeked shared a local statusline trick to read quota data from ~/.claude/settings.json without API calls (post), and several Show HN submissions (Claude Monitor, notch dashboard) build on this pattern. The pattern from April 15 continues: developers build workarounds for tool limitations rather than switching away.


5. What People Are Building

Project Who built it What it does Problem it solves Stack Stage Links
SPICE-MCP fizz_buzz MCP servers for oscilloscope + SPICE simulator integration with Claude Code Manual simulation-to-hardware verification loop Python, MCP, LeCroy SDK, spicelib Alpha post
Paper Lantern paperlantern MCP server surfacing techniques from 2M+ CS papers to coding agents Agents rely on training data, missing recent research MCP, npx Beta site, GitHub (post)
Egregore ohmyai Shared memory and coordination substrate for multiplayer Claude Code Teams using Claude Code diverge from common vision without shared context Claude Code hooks, git, Markdown Alpha GitHub (post)
AI Subroutines (rtrvr.ai) arjunchint Record browser tasks once, replay as deterministic scripts inside the tab Runtime AI browser agents are non-deterministic, expensive, auth-broken Browser extension, DOM/fetch interception Beta blog (post)
Co-op ajayarama 24/7 agent runner for non-technical users β€” email summaries, slide decks, finance tracking Non-technical users cannot run agents without laptops running 24/7 Mobile app, multi-service integrations Alpha site (post)
ShadowStrike Phantom Soocile Open-source EDR/XDR platform with AI/ML detection, kernel driver No open-source endpoint protection with full EDR/XDR capabilities C/C++, kernel driver, AI models Alpha GitHub (post)
Codg veni0 Multi-model async AI agent harness with TUI, web, desktop, and messaging modes Fragmented agent harness ecosystem, no local model support Go Beta GitHub (post)
mcp.hosting jeffyaw Cloud-synced MCP server configuration across Claude Code, Cursor, and other clients Hand-editing JSON config per client is annoying Fastify, Postgres, Caddy, EKS Shipped site (post)
Mabon luckystrike AI agent that continuously finds and matches jobs Job search requires constant manual checking Agent-based Alpha site (post)
Clamp sidneyottelohe Web analytics that AI agents can read and query Traditional analytics dashboards aren't agent-accessible Web Alpha site (post)
Agents.ml bayff Public identity page and A2A card for AI agents No standard way for agents to discover and identify each other Web Alpha site (post)
Vibe Games pzxc One vibe-coded video game per day using Claude Personal challenge to rebuild a games site Claude, web Shipped site (post)

The Show HN submissions cluster into three categories: (1) Claude Code ecosystem tools (Egregore, quota monitors, context engineering references), (2) agent infrastructure and identity (mcp.hosting, Agents.ml, Clamp, Codg), and (3) domain-specific agent applications (SPICE-MCP, Paper Lantern, Mabon, Oura Ring MCP). The Claude Code ecosystem category is the largest β€” 24 of 106 stories mention Claude Code directly, reflecting both the tool's dominance and its users' need to build around its limitations.

AI Subroutines from rtrvr.ai represents a maturing of the browser automation space. Their key architectural insight β€” executing scripts inside the webpage's own execution context so that "auth, CSRF, TLS session, and signed headers get added to all requests and propagate for free" β€” addresses the fundamental auth problem that plagues out-of-process browser agents. This continues the development-time browser automation trend identified on April 15 (Libretto).


6. New and Notable

Cloudflare Agent Memory: Persistent State for Agent Workers

tysont shared Cloudflare's introduction of persistent agent memory built on their Workers platform (post). Following April 15's Project Think announcement (durable execution for one-to-one agent sessions), this adds the persistence layer agents need to maintain state across interactions. The Cloudflare agent infrastructure stack is assembling rapidly: durable execution, sandboxed code, sub-agents, and now memory.

AI Agent Identity and Discovery Standards Proliferating

Three independent submissions on agent identity emerged in a single day: Agents.ml (public identity pages with A2A cards, post), AAIP (open protocol for AI agent identity and agent-to-agent commerce, post), and an open-source commitment protocol for AI agents (post). Combined with isitagentready.com's protocol discovery checks (MCP Server Card, Agent Skills, WebMCP, OAuth), the agent discovery and identity layer is crystallizing β€” though no standard has emerged.

Anthropic's Official Opus 4.7 Guidance: Delegate, Don't Pair

Anthropic's best practices post for Opus 4.7 introduced a notable framing shift: treat Claude "like a capable engineer you're delegating to" rather than "a pair programmer you're guiding line by line" (post). The new xhigh effort level (default, between high and max) and recommendation to batch questions and minimize user turns represent an explicit move toward autonomous agent behavior and away from the interactive coding assistant paradigm.

Perplexity Releases "Personal Computer"

MrBuddyCasino shared Perplexity's announcement of "Personal Computer," though the post received minimal engagement (3 points, 0 comments) (post). Notable as a product signal: search-native AI companies expanding into agent-like PC integration.

MCP Security Concerns Hit Critical Mass

beardyw shared a report that Anthropic won't acknowledge an MCP "design flaw" affecting 200K servers (post). ronxjansen separately argued that coding agents "degrade sandboxes to security theater" (post). These posts land the same week Amazon is publicly doubling down on MCP, creating a tension between adoption momentum and unresolved security architecture.


7. Where the Opportunities Are

[+++] Independent Model Cost Benchmarking β€” The tokenizer measurement post (480 points) demonstrated massive demand for honest, independent cost analysis that goes beyond vendor-published ranges. namnnumbr explicitly called for per-task cost analysis, not just per-token counts. As models iterate faster and pricing becomes harder to compare across vendors, an independent cost-intelligence service β€” measuring real-world cost-per-task across models, tokenizers, and effort levels β€” fills a growing trust gap. (post)

[+++] Agent-Resistant Open Source Tooling β€” Archestra.ai's experience (253 bot comments, 27 untested PRs, half-day weekly cleanup) and their crude workaround (contributor onboarding via git --author whitelisting) demonstrates that GitHub's current tooling cannot distinguish AI-generated contributions from human ones. A lightweight reputation/verification system for open-source repos β€” more sophisticated than CAPTCHAs, less nuclear than blocking all newcomers β€” is an urgent need as bounties and AI coding agents proliferate. (post)

[++] Structured Domain Adapters for Agent Workflows β€” The SPICE/oscilloscope discussion produced a clear architectural pattern: agents fail when reading domain-specific files directly but succeed when structured intermediaries (Python analyzers producing JSON) sit between the agent and the domain tool. This pattern generalizes to KiCad, EDA tools, CAD, and any domain with complex file formats. A library of domain-file-to-JSON adapters, optimized for agent consumption, would unlock hardware and engineering verticals. (post)

[++] Research-Augmented Agent Pipelines β€” Paper Lantern's benchmark showing 30-80% improvement on 5 of 9 coding tasks by surfacing recent research papers suggests a viable product category. Two independent builders (Paper Lantern and vunderba's Go-based paper-search) arrived at the same insight. The key finding: "10 of 15 most-cited papers across all experiments were published in 2025 or later," meaning training data alone cannot substitute for current research. (post)

[+] Agent Identity and Discovery Layer β€” Three independent submissions on agent identity in one day (Agents.ml, AAIP, commitment protocol) plus isitagentready.com's protocol checks signal the "DNS for agents" problem becoming urgent. No standard has won, creating an opening for whoever builds the simplest, most adopted agent card format. (post, post)

[+] Team Coordination for Multi-Agent Coding β€” Egregore's approach (shared git-backed memory, /handoff, /invite, Claude Code hooks) addresses the divergence problem when multiple people use Claude Code on the same codebase. Combined with April 15's agent observability tools (Jeeves, Lazyagent), the need for team-scale agent coordination is growing beyond individual session management. (post)


8. Takeaways

  1. Opus 4.7's tokenizer inflates real-world costs more than Anthropic disclosed. Independent measurement showed 1.325x weighted across typical Claude Code content, with technical docs at 1.47x β€” above Anthropic's stated 1.0-1.35x range. Same price, faster quota burn, shorter effective context window. (post)

  2. The Claude developer experience is being quietly degraded, and the community is documenting it. Removed extended thinking budgets, cache TTL cuts, third-party token blocking, and usage policy false positives form a pattern that one blogger called "adversarial." Anthropic's own response β€” recommending fewer interactive turns and a new effort level β€” tacitly confirms the capacity constraints. (post, post)

  3. Agent readiness for the web is being pushed by infrastructure players and resisted by developers. Cloudflare's scanner and Amazon's MCP investment signal enterprise adoption, but HN commenters overwhelmingly want agent-blocking tools, not agent-enabling ones. The SEO-to-GEO pivot analogy resonated. (post, post)

  4. Hardware and engineering domains expose a fundamental agent limitation: no physics intuition. SPICE, KiCad, and circuit design experiences converged on the same conclusion β€” agents hallucinate when reading domain files directly but work when structured JSON intermediaries translate the domain. (post)

  5. AI-generated open source contributions have become an operational cost. Archestra.ai's 253-comment bot invasion and "nuclear option" of contributor gating is the clearest case study yet. GitHub's existing tools are insufficient for the AI spam problem. (post)

  6. Research-augmented coding agents outperform baseline agents by 30-80% on technique-sensitive tasks. Paper Lantern's benchmarks and an independent similar system suggest that connecting agents to current research literature is a genuine capability multiplier, not just a demo. (post)

  7. Agent identity, discovery, and coordination standards are proliferating without convergence. Three independent agent identity projects in one day plus MCP, WebMCP, Agent Skills, A2A, and AAIP create a fragmented landscape. The winner will be whoever achieves the simplest adoption path β€” not the most comprehensive spec. (post, post, post)