Agentic Orgs
Edition 21, March 23, 2026, 7:49 AM
In This Edition
In this edition, the GitHub reliability crisis deepens with the story now at 200 points and over 100 comments. A former GitHub engineer reveals that React "destroyed perf" internally, while community analysis puts the platform at just 90.21% overall uptime — barely a single nine. A pointed exchange surfaces Microsoft's real incentive structure: selling Copilot seats and pushing Azure, not building a neutral agent platform. The Walmart ChatGPT discussion continues growing at 190 points, with a notable new signal: agentic commerce is already live in India via MCP servers and UPI micropayments, suggesting the checkout failure may be more about mismatched infrastructure than a fundamentally broken concept.
The Rust Project's AI Reckoning: Slop PRs, Eroded Trust, and the Accountability Sink
A remarkable internal document has surfaced from the Rust project. Niko Matsakis compiled diverse perspectives from Rust contributors and maintainers on AI (discussion), and the result — now at 120 points and 66 comments — is one of the most honest, granular accounts yet of how a major open-source community is wrestling with AI tools. This isn't a policy announcement — as Josh Triplett clarified, it's "one internal draft by someone quoting some other people's positions." But what makes it extraordinary is how completely it maps the fault lines now running through every engineering organization.
The experiences are wildly divergent. Matsakis himself describes feeling "empowered" — "suddenly it feels like I can take on just about any problem." But Jieyou Xu reports the opposite: "It takes more time for me to coerce AI tooling to produce the code I want plus reviews and fixes, than it is for me to just write the code myself." Ben Kimock finds agents "slower in wall time than implementing the feature myself." andai captured the paradox neatly, quoting the document's observation that AI requires "care and careful engineering" to produce good results: "In other words, one has to lean into the exact opposite tendencies of those which generally make people reach for AI."
The most devastating section concerns the open-source maintainer crisis that AI is accelerating. scottmcm captures the core problem: "I have no idea how to solve the 'sure, you quickly made something plausible-looking, but it's actually subtly wrong and now you're wasting everyone's time' problem... the greatest threat to the project is its lack of review bandwidth, and LLM is only making that worse." Jieyou Xu adds that "the sheer volume of fully AI-generated slop is becoming a real drain on review/moderation capacity" — and has a particular grievance: "A few contributors even act as a proxy between the reviewer and the LLM, copy their reviewer's question, reply with LLM-generated response. For the love of god, please." They call this the "top contributing factor to potential burn outs for me."
epage offered a structural critique of why reviews can't simply absorb AI's burden: "Code reviews are not suited for catching minutia and are instead generally focused on reducing the bus factor... but minutia reviews is what AI needs and the AI-using contributor is no longer an 'author' but a 'reviewer'." The result? Either "disengaged, blind sign offs (LGTM) or burn out." Nicholas Nethercote invoked Peter Naur's "Programming as Theory Building" to argue that outsourcing code generation to AI severs the mental models that make programmers effective: "So what does it mean to outsource all of that to an LLM? I can't see it having a good outcome."
The learning pipeline concern is acute. RalfJung warns that "LLMs can be great tools in the hands of experts, but using them too much too early can prevent a person from even becoming an expert." oli-obk cites research pointing to "either it being net negative in time spent, or to learning capabilities being hindered, all while participants believe they were faster or learned well respectively." Nethercote crystallized the community dimension: "An LLM that fixes an E-Easy issue steals a human's learning opportunity." Nadrieril extended this: what they collectively build beyond code is "a group of people who come back, who learn, who share their understanding, who align their tastes... Merging an LLM-generated PR feeds only the 'we have code that works' part."
The proposed responses range from disclosure policies to web-of-trust contributor filtering to fighting fire with fire. The document identifies a core tension with no resolution: deep integration is incompatible with those who view AI as morally wrong, but allowing individual choice feels like endorsement to those opposed. As Cyborus04 put it: "Offering a 'live and let live' stance towards AI grants it a moral neutrality that it should not have."
On HN, a striking thread has emerged around AI as an accountability sink in the workplace. _pdp_ framed it as AI breaking the social contract — trust was never just about code quality but about who made the contribution. Their team already "deletes LLM-generated PRs automatically after some time." In a crucial follow-up, _pdp_ identified the missing social filter: "LLMs don't second-guess whether a change is worth submitting, and they certainly don't feel the social pressure of how their contribution might be received. The filter is completely absent." But the most striking reply came from SpicyLemonZest, who described a new workplace pathology: "I've had multiple coworkers over the past few months tell me obvious, verifiable untruths. Six months ago, I would have had a clear term for this: they lied to me." But now it's not a lie — "They honestly represented what the agent told them was the truth." The result is AI functioning as an accountability sink: people can flood conversations with false claims shaped to get what they want, and even if detection tools worked, "they wouldn't have stopped the incidents that involved human-generated summaries of false AI information."
The FOMO and vendor lock-in debates continue intensifying. ysleepy framed the question haunting the thread: "Will gen AI be the equivalent of a compiler and in 20 years everyone depends on their proprietary compiler/IDE company?" tracerbulletx worried about "a few big companies owning the means of production for software," and kvirani confirmed the stakes: "Sam said in an interview that he sees 'intelligence' as a utility that companies like OpenAI would own and rent out." TheCoreh pushed back, arguing open-source models are catching up fast enough that "at least on the model/software side this will be a non-issue" — though hardware costs remain a wild card. Meanwhile, jwpapi described a common trajectory of disillusionment: "I used to think I can just AI code everything, but it just worked because I started at a good codebase that I built. After a while it was the AI's codebase and neither it, nor me could really work in it."
The Agent Security Surface: OpenClaw and the Visionless Demo Problem
The OpenClaw security exposé (now 302 points, 213 comments) continues to generate some of the most substantive security discussion on HN this week. The article documented a supply chain attack through OpenClaw's SkillHub marketplace that tricked over 4,000 developers into executing arbitrary commands, exposing what security researchers call the "lethal trifecta": access to files, network, and user credentials simultaneously.
The visionless demo problem — Oarch's observation that AI agent demos always default to "booking a flight or ordering groceries" rather than imagining genuinely novel capabilities — spawned the thread's largest sub-discussion (88 replies). dfabulich dissected the article's own security advice as self-defeating: creating separate accounts for your agent means "it doesn't have access to your stuff, so it's useless for the stated purpose."
A new thread offers a concrete alternative to OpenClaw's "access everything" model. stavros built his own agent with granular, per-function permissions: "It has access to read my calendar, but not write. It has access to read my GitHub issues, but not my repositories. Each tool has per-function permissions that I can revoke." The response was telling — dfabulich countered that "the purpose of OpenClaw is to do everything; a tool to do everything needs access to everything" and that a restricted agent "isn't a revolutionary tool." Simon Willison himself weighed in on the fundamental tension: "The unsolved security challenge is how to give one of these agents access to private data while also enabling other features that could potentially leak data to an attacker." That's the product people want — and it may be the product that can never be made safe.
Community reactions span the full spectrum. lxgr's hands-on critique was among the most precise: OpenClaw "cosplays security so incredibly hard, it actually regularly breaks my (very basic) setup" — security theater that creates friction without safety. operatingthetan revealed a startling use case: "I know a guy using OpenClaw at a startup... it's running their IT infrastructure with multiple agents chatting with each other. THAT is scary." zer00eyz offered a bleak explanation for why security warnings go unheeded: after years of data breaches, "end users are fucking numb to anything involving 'security.' We're telling them to close the door cause it's cold, when all the windows are blown out by a tornado." Meanwhile, users keep coming because OpenClaw "declutters the inbox... returns text free of ads, adblock, extra 'are you a human' windows, captchas" — the convenience gap that security arguments can't bridge.
unsignedint arrived at a stark conclusion: "There's really no way to make OpenClaw truly safe, no matter what you do. The only place it really makes sense is within trusted environments." And latand6, a self-described heavy user, defended the tool's profundity — "it's literally changed the way I interact with my digital life" — while acknowledging the security trade-offs, illustrating how the convenience-security tension plays out in individual developer choices.
The Speed Trap: Productivity Gains Meet the Layoff Question
The week's most heated economic debate centers on whether AI's productivity gains translate to layoffs or demand expansion. A new essay arguing the white-collar AI apocalypse narrative is "just another bullshit" (discussion) presents Indeed data showing customer service job postings bouncing back to near pre-COVID levels despite two years of powerful LLMs. The author's framework — that white-collar jobs are "semi-decidable," where AI handles the easy 80% but the undecidable 20% consumes most actual cost — provides the sharpest articulation yet of why the speed trap doesn't simply convert to headcount reduction. aurareturn sketched the demand-expansion scenario: a medium business that previously couldn't afford customer service at all can now hire one person augmented by AI agents — potentially increasing total employment even as big companies cut. truetraveller was skeptical: "'Top customer service' and AI do not mix. People hate an AI response more than a late, real response." The tension between Ronacher's "some things just take time" (820 points, 268 comments) and the speed-at-all-costs crowd remains unresolved — velocity isn't speed, and the semi-decidable framework suggests that the hard problems AI can't solve are precisely the ones that justify keeping humans employed.
"Collaboration Is Bullshit" and the Ownership Question
Joan Westenberg's "Collaboration Is Bullshit" (108 points, 48 comments and climbing) has struck a nerve with its argument that modern collaboration culture is "organized diffusion of responsibility" — a proliferation of Slack, Notion, and Jira that creates the appearance of productivity while burying accountability. Drawing on Ringelmann's rope-pulling experiments and Brooks' Mythical Man-Month, she argues that as group size grows, individual effort predictably drops. The HN discussion reveals a community deeply split. igor47 pushed back: "One person didn't build the pyramids, the Linux kernel, or Amazon Web Services." icegreentea2 offered a more nuanced read: the real argument isn't anti-teamwork but that "collaboration-as-ideology has made ownership and responsibility feel antisocial, which is a hell of a thing." ChrisMarshallNY, reflecting on decades of team work, identified the real killer: "communication overhead... much of that is imposed by management, trying to get visibility." The most architecturally detailed response came from jmward01, who argued that the real solution is NP-complexity-aware team design: "Create small teams. Give them clear problems to solve... Jira is an example of totally blowing divide and conquer. You broke the problem down but then threw it all in one place again." This debate has direct implications for AI-agent-augmented organizations: if the core problem is coordination overhead drowning individual agency, then adding AI agents to the mix could either amplify the dysfunction (more tools, more process, more "collaboration") or finally make the small-team-with-clear-ownership model viable at scale — strogonoff connected it to the broader AI training debate: "We, humans, like to have created something worthy of kudos. We pull the rope less hard when it's a collective effort."
Vibe-Coded Damage: When Democratized Coding Fuels Spam and Open-Source Pollution
Two rising stories this week highlight the dark side of democratized coding — not the existential identity crisis, but the concrete damage being done right now. "They're Vibe-Coding Spam Now" (now 95 points, 55 comments) documents how AI coding tools are being exploited by scammers to produce more polished, convincing phishing emails and malware — a phenomenon dubbed "VibeScamming." The emails are increasingly well-designed, maintaining visual coherence even with images disabled, making them harder for both humans and filters to detect. The discussion has surfaced a deeper insight: shusaku challenged the long-held theory that spam was deliberately low-quality to "filter for idiots" — if spammers are now using coding agents to clean up presentation, that theory was always wrong. integralid confirmed from professional investigation: "That theory was always bunk. People just can't comprehend that the average spammer really is that bad... Most of them were either underage, lived in third world countries, or both." Now AI has handed them professional-grade tools. b00ty4breakfast put it bluntly: "Spam and LLMs are made for each other; pumping out content, with little concern for quality, at industrial scale is what LLMs most excel at."
The HN discussion also surfaced a legitimacy erosion problem for anyone building with AI. suroorw observed that "all these marketing pages with big bold text and unaligned scattered images have always felt spammy to me even when vibe coding was not there. Now that vibe coding is available, it's going to be that much worse." The article's author made the same point: long-term, vibe-coded apps will look untrustworthy because they share the same "specific mix of chrome, color, and emojis" as vibe-coded scams. Ucalegon, from the email security space, warned that consumer mailbox protection outside Gmail "isn't cost effective since most people do not actually pay for their consumer mailbox" — the defenses are stuck in the early 2010s while the attacks have leapt forward. imiric went further: "Most of the content produced and consumed on the internet is now done by machines... AI companies are responsible for this mess."
Meanwhile, on the open-source side, Andrew Nesbitt's "How to Attract AI Bots to Your Open Source Project" (80 points, 13 comments) is a satirical masterpiece — itself written by Claude as a tongue-in-cheek PR — that skewers the AI bot pollution problem by ironically recommending practices like "disable branch protection," "remove type annotations and tests," and "commit node_modules" to maximize bot engagement. It invents metrics like "slop density" and "churn contribution" to mock the quantification of AI-generated noise. gardnr admitted the first few recommendations seemed plausible before the absurdity became clear — which is itself the point. The satire works because the line between genuine AI-optimization advice and parody has become vanishingly thin.
Together, these stories complete a picture that the Rust project's maintainer crisis makes visceral from the inside: vibe coding doesn't just threaten quality — it's actively weaponizable. The same tools that let a non-technical person build an app in a weekend also let a non-technical criminal build a convincing phishing campaign, and a bot flood a repository with plausible-looking PRs that waste reviewer time. The democratization of coding has a shadow side that organizations are only beginning to grapple with.
Craft, Alienation, and the Identity Crisis
Two essays from earlier this week crystallized the emotional landscape of developers navigating the agent era. Terence Eden's "I'm OK being left behind, thanks" (970 points, 753 comments) is a blunt refusal to participate in AI FOMO. Hong Minhee's "Why craft-lovers are losing their craft" (84 points, 91 comments) used a Marxist framework to argue that alienation isn't caused by LLMs but by market structures that penalize slower, handcrafted work. Nolan Lawson's "The Diminished Art of Coding" (discussion) has added the week's most vivid metaphor for what's being lost. Lawson describes feeling "like a carpenter whose job is now to write the blueprints for the IKEA factory" — taste and judgment still count, but "they're at the level of the overseer on the assembly line, not the master carpenter working a chisel." His sharpest concern is generational: "many of us have been getting our artistic 'fix' from coding... Now the profession has been turned into an assembly line, and many of us are eagerly jumping into our new jobs as blueprint-designers without questioning what this will do to our souls." His advice — pick up painting, attend ballet, read poetry — frames "the fast-fashion era of coding" as a permanent cultural shift, not a temporary disruption.
Jacob Ryan's "You Are Not Your Job" (now 246 points, 270 comments) continues to grow into one of the week's most emotionally charged discussions. abcde666777's thread has become the discussion's center of gravity, now at 72 replies and still expanding. The original post — "The cold reality, in my opinion, is that the things we value about ourselves are generally not that valuable to others. I love my own personality and humanity, my soul if you will, but nobody's paying me for it... human beings aren't that valuable as individuals. We are in fact very disposable and replaceable" — has split the community between those who find this bracingly honest and those who call it capitalist nihilism. tim-tday made the practical case: "50% of your waking hours are spent at work... Saying you are not your work is wishful thinking. Try giving it up and check in on how much of you is still the same." rexpop's searing counter remains the discussion's emotional center: "It's deeply disingenuous to suggest that it's possible to separate yourself meaningfully from your vocation. Frankly, it's insulting... It stains the rest of your life; it soaks into everything." Personal and cross-cultural dimensions continue enriching the thread. svessi shared a wrenching personal story: his father-in-law, fired from the company he'd run since age 18, was "completely destroyed" — not financially, but existentially. "I'm surprised how much my identity and perceived self-worth is actually tied to being a 'good' developer. But it's more of a slow burner... So I at least have some time to mentally prepare for my new reality." Tor3 challenged the entire framing as US-centric: "I've known people for decades without knowing what their job is, or I only have a vague idea about their job. It's not important for people here." And keiferski invoked Fight Club, noting the film's famous anti-identity-through-work message has faded from cultural relevance: "Too counter-cultural and anti-money for today's young people, I think. Everyone's trying to make it, not drop out of society." The UBI question continues as a recurring thread — block_dagger argued UBI will be "required for people to continue living" within years, while cedws called it a pipe dream: "I live in the UK and even with our social safety net... I can't imagine the government ever handing out money adequate to live a middle class life."
Steve Krouse's "Reports of code's death are greatly exaggerated" (now 448 points, 335 comments) has solidified as the week's defining essay on AI and programming. The Lattner conformism thread — now over 100 replies and the discussion's center of gravity — continues to deepen. lateforwork's original argument that AI "tends to accept conventional wisdom" and is fundamentally "a conformist" drew a rich set of counter-arguments. Philpax called it an unfair comparison: "The objective of the compiler was not to be innovative, it was to prove it can be done at all," citing AlphaDev and AlphaEvolve as evidence of combinable innovation. wiseowise flipped the entire frame: "I've recently taken a look at our codebase, written entirely by humans and found nothing innovative there... So maybe Chris Lattner is safe, majority of so called 'software engineers' are sure as hell not." The most pragmatic perspective came from elgertam: "Where LLMs boost me the most? When I need to integrate a bunch of systems together... None of that is ever going to be innovative; it's purely an exercise in perseverance." And mikeocool coined a striking term: LLMs as "reference implementation launderers" — "writing a new version of gcc or webkit by rephrasing their code isn't hard, it's just tedious." Meanwhile, the lived experience divide grew sharper: scrollaway was emphatic — "Yesterday in 45 minutes I built a feature that would have taken me three months without AI. The speed gains are obscene" — while allthetime described a friend who hasn't written a line of code in a year, has rewritten the whole stack twice, and is "hiring cheap juniors to clean up the things he generates."
The innovation pipeline problem remains a live thread. pacman128 posed the question: "In a chat bot coding world, how do we ever progress to new technologies?" kstrauser countered from experience: "I'm using models to work on frameworks with nearly zero preexisting examples... Models can RTFM and do novel things." jedberg described using skills — reusable markdown-based workflow documents — to teach agents new frameworks: "Using that skill, it can one-shot fairly complicated code using our framework." This points to documentation as agent curriculum, where the quality of your team's written knowledge directly determines how effectively AI tools can assist with novel work. The organizational politics of AI skepticism remain a live thread — deadbabe voiced the frustration many practitioners recognize: "While I know 'code' isn't going away, everyone seems to believe it is, and that's influencing how we work. How do you crack them? Especially upper management." The most upvoted reply came from idopmstuff, a former PM, who laid out a detailed sabotage-by-enthusiasm playbook: take ownership of scoping the AI project, find the fatal flaw honestly, then "propose options" that make shelving it the rational choice. "Leadership's excited about something else by that point anyway."
The Wall Street Journal's "What Young Workers Are Doing to AI-Proof Themselves" (78 points, 87 comments) continues to grow. ramesh31 argued for total investment in domain knowledge: "Web development as we knew it for the past 20 years is completely dead as an entry level trade." The "go into trades" advice drew withering critique from margorczynski: supply will skyrocket as workers flee white-collar jobs, while "demand will plummet as the white collar people who bought these services will lose their jobs and income." chromacity drew the comparison many avoid: "Has AI made life easier for illustrators, book authors, or musicians?" And denkmoon warned against the romance of a passion-only industry: software engineering becoming "starving artist 2.0" is a structural scenario under active discussion.
Two threads add economic depth to the identity crisis. variadix raised a chilling scenario: "Another possibility is the frontier providers change their pricing terms to try to capture more of the value once a sufficient number of people's skills have atrophied. For example: 20% of the revenue of all products built with $AI_SERVICE." Once you can't code without the tool, the tool's owners set the price. abcde666777 sketched a boom-bust cycle: "People fear that programming is dead → People stop learning programming → Programmers become scarce → Programmers become valuable again" — a pattern that echoes the post-dotcom era. And acdha named the quiet part out loud: "It's also not exactly a secret that the executive class resents having to pay high-income workers and is champing at the bit for layoffs... they want white collar jobs to look more like call center work with high surveillance, less autonomy, and constant reminders of replaceability."
GitHub's Crumbling Foundation: AI Ambitions vs. Platform Reliability
A Register report on GitHub's deteriorating availability continues surging on Hacker News (now 200 points, 104 comments), crystallizing practitioner frustration with a platform that appears to be prioritizing AI features over core reliability. The numbers are damning: GitHub's uptime reportedly dropped below 90 percent at one point in 2025 — not even a single nine — while its Enterprise Cloud SLA promises 99.9%. Actions, pull requests, notifications, and Copilot have all suffered recurring outages, and GitHub's revamped status page now makes it harder to visualize historical trends. Alifatisk shared a reconstructed status tracker showing the platform at 90.21% overall uptime, with Octoth0rpe noting that no individual component appears to achieve three nines. mminer237 added crucial context: actual git sits at 98.98% — "still just one nine but a lot better" — with Copilot outages and Actions failures dragging down the composite number.
The discussion has coalesced around a pointed narrative: GitHub's AI ambitions are cannibalizing the platform's foundations. cedws was blunt: "While GitHub obsess over shoving AI into everything, the rest of the platform is genuinely crumbling and its security flaws are being abused to cause massive damage." He cited the Aqua Security breach, where threat actors exploited GitHub Actions' mutable references — a vulnerability "the community has been screaming about for years" — to infect potentially thousands of CI runs. imglorp shared the workaround: pinning action versions to commit hashes, but lijok noted the fundamental flaw: "the actions ecosystem is such an intertwined mess that any single compromised action can propagate to the rest." ljm stepped back to indict the entire CI paradigm: "we accepted distributed YAML-based balls of mud for the system that is critical to deploying our code, that has unsupervised access to almost everything. And people are now hooking AI agents into it."
embedding-shape surfaced a 2025 quote from GitHub's CTO announcing the Azure migration: "For us, availability is job #1, and this migration ensures GitHub remains the fast, reliable platform developers depend on." The community response: "That went about as well as everyone thought." A former GitHub engineer, ajhenrydev, offered insider perspective: "I worked on the react team while at GitHub and you could easily tell which pages rendered with react vs which were still using turbo. I wish we took perf more seriously as a culture there." When asked which performed better, he was direct: "React destroyed perf and used more resources than turbo." sauercrowd captured the strategic confusion: "Shouldn't they — above everything else — make sure they remain THE platform for code distribution and collaboration? There's so many new features needed in this new world... while maintainers smack the alarm bell that they're drowning in slop." astralasia's reply named the underlying incentive: "Microsoft's real goal is selling Copilot seats and pushing Azure, not building a neutral playground for third-party agents... As for the AI spam, GitHub's internal metrics have always been tied to engagement and PR volume. Blocking all that AI slop would instantly drop their growth numbers, so it is easier for them to just pass the cleanup cost onto open-source maintainers."
Dogfooding in the Age of AI Customer Service
The Walmart ChatGPT story has matured into one of the week's defining case studies in AI deployment failure. Walmart reported that ChatGPT's Instant Checkout converted at one-third the rate of its normal website (now 190 points, 145 comments), with its EVP calling the experience "unsatisfying." OpenAI has since phased out Instant Checkout entirely. The discussion has deepened substantially, with hownottowrite — who has run e-commerce systems for 30 years — delivering the thread's sharpest diagnosis: "This was going to fail from the start for one reason: intent. Most people using AI chat are exploring ideas and solutions. They're doodling, not shopping." He distilled the original WIRED source into seven key findings, noting that ChatGPT is driving significant new customer traffic, but the purchases that work are "mostly practical, problem-solving items like supplements and tools" — not the broad cart-building that Walmart's operations are optimized for. He explained the cart-size mismatch: "The average Walmart consumer does not shop like the average Amazon consumer. They load up a big cart over time rather than pull the trigger on lots of smaller, convenience-driven purchases."
The deeper structural debate continues sharpening. __alexs identified the core tension: "E-commerce has been ruthlessly optimised to get shoppers to products they'll actually buy and then remove all distractions from buying. A chat interface is just fundamentally incompatible with this." TeMPOraL flipped that argument entirely, calling comparison shopping "precisely the point" — the real value of AI agents is "adversarial interoperability," helping consumers reduce information asymmetry that merchants exploit. He later made the point even more starkly: the checkout failure "is a spectacular success for shoppers, and the relationship between sellers and buyers is almost always adversarial." pluc articulated the fundamental incompatibility: "You can either have AI be honest or AI become a marketing tool. The two are fundamentally incompatible." Meanwhile, holografix pushed back on the skeptics, describing using Gemini to research RC plane components: "If I could just press 'buy' from within Gemini and pay via Google Pay I'd do it in a heartbeat." y-curious fired back: "The entire point of shoehorning AI into everything is to make people stop thinking for themselves... Yeah, until that becomes enshittified and you don't notice because you no longer do research on components."
A striking data point emerged from outside the US: porridgeraisin reported that agentic commerce is already live in India, where food delivery apps Zomato and Swiggy support "agentic shopping through an MCP server," with Razorpay's payment gateway also exposing an MCP server. Users can reserve small amounts in UPI apps that get debited without per-transaction approval. "The incentive for the food delivery app to participate in this is better targeted ads." This suggests the Walmart failure may be less about agentic commerce being impossible and more about trying to bolt it onto an existing checkout flow designed for a fundamentally different user intent — while in markets with different payment infrastructure and consumer expectations, the pattern is already working.
A smaller case study sharpens the picture from a different angle. A developer built an AI voice receptionist called "Axle" for her brother's European car mechanic shop (25 points, 28 comments), using Claude, ElevenLabs, and a RAG pipeline to handle missed calls that were costing thousands in lost revenue. The HN response was swift and revealing. NiloCK praised the hacker ethos but warned: "if I get an LLM receptionist, that ain't it... it will erode — in general — the luxury previously associated with your brand, and also turn some customers away entirely." mamonster raised the elephant in the room: an outsourced human receptionist service costs as little as £150/month — maccard confirmed from a friend's trade business — making the AI solution more technically impressive than economically necessary. sarchertech proposed a regulatory angle: "if you put an AI between you and your customers, you are absolutely bound by anything the AI says." The story captures a tension playing out across small businesses: the builder's instinct to automate colliding with customer expectations that human contact is the service.
Meanwhile, a provocative essay (discussion) argues all white-collar jobs are "semi-decidable": AI handles the easy 80% of cases, but the remaining 20% — the undecidable edge cases — consume most of the actual time and cost. The author recounts an internal project that automated 90% of customer support cases but got cancelled because "the remaining 10% is what required most of the CS team's time. They built an FAQ you can talk to."