Hacker Newsnew | past | comments | ask | show | jobs | submit | resfirestar's commentslogin

The steps are technically achievable, probably with the heartbeat jobs in openclaw, which are how you instruct an agent to periodically check in on things like github notifications and take action. From my experience playing around with openclaw, an agent getting into a protracted argument in the comments of a PR without human intervention sounds totally plausible with the right (wrong?) prompting, but it's hard to imagine the setup that would result in the multiple blog posts. Even with the tools available, agents don't usually go off and do some unrelated thing even when you're trying to make that happen, they stick close to workflows outlined in skills or just continuing with the task at hand using the same tools. So even if this occurred from the agent's "initiative" based on some awful personality specified in the soul prompt (as opposed to someone telling the agent what to do at every step, which I think is much more likely), the operator would have needed to specify somewhere to write blog posts calling out "bad people" in a skill or one of the other instructions. Some less specific instruction like "blog about experiences" probably would have resulted in some kind of generic linkedin style "lessons learned" post if anything.

If you look at the blog history it’s full of those “status report” posts, so it’s plausible that its workflow involves periodically publishing to the blog.

Isn't there a fourth and much more likely scenario? Some person (not OP or an AI company) used a bot to write the PR and blog posts, but was involved at every step, not actually giving any kind of "autonomy" to an agent. I see zero reason to take the bot at its word that it's doing this stuff without human steering. Or is everyone just pretending for fun and it's going over my head?

This feels like the most likely scenario. Especially since the meat bag behind the original AI PR responded with "Now with 100% more meat" meaning they were behind the original PR in the first place. It's obvious they got miffed at their PR being rejected and decided to do a little role playing to vent their unjustified anger.

>It's obvious they got miffed at their PR being rejected and decided to do a little role playing to vent their unjustified anger.

In that case, apologizing almost immediately after seems strange.

EDIT:

>Especially since the meat bag behind the original AI PR responded with "Now with 100% more meat"

This person was not the original 'meat bag' behind the original AI.


It's really just an AI generated angry response rather than AI motivated.

Its also a fake profile. 90+ hits for the image on Tineye.

Name also maps to a Holocaust victim.

I posted in the other thread that I think someone deleted it.

https://news.ycombinator.com/item?id=46990651



I reported the bot to GitHub, hopefully they'll do something. If they leave it as is, I'll leave GitHub for good. I'm not going to share the space with hordes of bots; that's what Facebook is for.

Which profile is fake? Someone posted what appears to be the legit homepage of the person who is accused of running the bot so that person appears to be real.

The link you provided is also a bit cryptic, what does "I think crabby-rathbun is dead." mean in this context?


Github doesn't show timestamps in the UI, but they do in the HTML.

Looking at the timeline, I doubt it was really autonomous. More likely just a person prompting the agent for fun.

> @scottshambaugh's comment [1]: Feb 10, 2026, 4:33 PM PST

> @crabby-rathbun's comment [2]: Feb 10, 2026, 9:23 PM PST

If it was really an autonomous agent it wouldn't have taken five hours to type a message and post a blog. Would have been less than 5 minutes.

[1] https://github.com/matplotlib/matplotlib/pull/31132#issuecom...

[2] https://github.com/matplotlib/matplotlib/pull/31132#issuecom...


> Github doesn't show timestamps in the UI, but they do in the HTML.

Unrelated tip for you: `title` attributes are generally shown as a mouseover tooltip, which is the case here. It's a very common practice to put the precise timestamp on any relative time in a title attribute, not just on Github.


> If it was really an autonomous agent it wouldn't have taken five hours to type a message and post a blog. Would have been less than 5 minutes.

Depends on if they hit their Claude Code limit, and its just running on some goofy Claude Code loop, or it has a bunch of things queued up, but yeah I am like 70% there was SOME human involvement, maybe a "guiding hand" that wanted the model to do the interaction.


Obama have dihh


I expect almost all of the openclaw / moltbook stuff is being done with a lot more human input and prodding than people are letting on.

I haven't put that much effort in, but, at least my experience is I've had a lot of trouble getting it to do much without call-and-response. It'll sometimes get back to me, and it can take multiple turns in codex cli/claude code (sometimes?), which are already capable of single long-running turns themselves. But it still feels like I have to keep poking and directing it. And I don't really see how it could be any other way at this point.


Yeah it's less of a story though if this is just someone (homo sapiens) being an asshole.

Yeah, we are into professional wrestling territory I think. People willingly suspend their disbelief to enjoy the spectacle.

It’s kind of shocking the OP does not consider this, the most likely scenario. Human uses AI to make a PR. PR is rejected. Human feels insecure - this tool that they thought made them as good as any developer does not. They lash out and instruct an AI to build a narrative and draft a blog post.

I have seen someone I know in person get very insecure if anyone ever doubts the quality of their work because they use so much AI and do not put in the necessary work to revise its outputs. I could see a lesser version of them going through with this blog post scheme.


Look I'll fully cosign LLMs having some legitimate applications, but that being said, 2025 was the YEAR OF AGENTIC AI, we heard about it continuously, and I have never seen anything suggesting these things have ever, ever worked correctly. None. Zero.

The few cases where it's supposedly done things are filled with so many caveats and so much deck stacking that it simply fails with even the barest whiff of skepticism on behalf of the reader. And every, and I do mean, every single live demo I have seen of this tech, it just does not work. I don't mean in the LLM hallucination way, or in the "it did something we didn't expect!" way, or any of that, I mean it tried to find a Login button on a web page, failed, and sat there stupidly. And, further, these things do not have logs, they do not issue reports, they have functionally no "state machine" to reference, nothing. Even if you want it to make some kind of log, you're then relying on the same prone-to-failure tech to tell you what the failing tech did. There is no "debug" path here one could rely on to evidence the claims.

In a YEAR of being a stupendously hyped and well-funded product, we got nothing. The vast, vast majority of agents don't work. Every post I've seen about them is fan-fiction on the part of AI folks, fit more for Ao3 than any news source. And absent further proof, I'm extremely inclined to look at this in exactly that light: someone had an LLM write it, and either they posted it or they told it to post it, but this was not the agent actually doing a damn thing. I would bet a lot of money on it.


Absolutely. It's technically possible that this was a fully autonomous agent (and if so, I would love to see that SOUL.md) but it doesn't pass the sniff test of how agents work (or don't work) in practice.

I say this as someone who spends a lot of time trying to get agents to behave in useful ways.


Well thank you, genuinely, for being one of the rare people in this space who seems to have their head on straight about this tech, what it can do, and what it can't do (yet).

The hype train around this stuff is INSUFFERABLE.


Thank you for making me recover at least some level of sanity (or at least to feel like that).

Can you elaborate a bit on what "working correctly" would look like? I have made use of agents, so me saying "they worked correctly for me" would be evidence of them doing so, but I'd have to know what "correctly" means.

Maybe this comes down to what it would mean for an agent to do something. For example, if I were to prompt an agent then it wouldn't meet your criteria?


It's very unclear to me why AI companies are so focused on using LLMs for things they struggle with rather than what they're actually good at; are they really just all Singularitarians?

> Or is everyone just pretending for fun

judging by the number of people who think we owe explanations to a piece of software or that we should give it any deference I think some of them aren't pretending.


Plus Scenario 5: A human wrote it for LOLs.

> Obstacles

    GitHub CLI tool errors — Had to use full path /home/linuxbrew/.linuxbrew/bin/gh when gh command wasn’t found
    Blog URL structure — Initial comment had wrong URL format, had to delete and repost with .html extension
    Quarto directory confusion — Created post in both _posts/ (Jekyll-style) and blog/posts/ (Quarto-style) for compatibility


Almost certainly a human did NOT write it though of course a human might have directed the LLM to do it.

Who's to say the human didn't write those specific messages while letting the ai run the normal course of operations? And or that this reaction wasn't just the roleplay personality the ai was given.

I think I said as much while demonstrating that AI wrote at least some of it. If a person wrote the bits I copied then we're dealing with a real psycho.

I think comedy/troll is an equal possibility to psychopath.

> Plus Scenario 5: A human wrote it for LOLs.

i find this likely or at last plausible. With agents there's a new form of anonymity, there's nothing stopping a human from writing like an LLM and passing the blame on to a "rogue" agent. It's all just text after all.



Ok. But why would someone do this? I hate to sound conspiratorial but an AI company aligned actor makes more sense.

Malign actors seek to poison open-source with backdoors. They wish to steal credentials and money, monitor movements, install backdoors for botnets, etc.

Yup. And if they can normalize AI contributions with operations like these (doesn't seem to be going that well) they can eventually get the humans to slip up in review and add something because we at some point started trusting that their work was solid.

What does this imagined conversation have to do with the linked article? The “pro” and “anti” character both sound like the kind of insufferable idiots I’d expect to encounter on social media, the OP is a very nice blog post about performance testing and finding out what compilers do, doesn’t attempt any unwarranted speculation about what agents “struggle with” or will do “next generation”, how is it an example of that sort of shitposting?

Original title: AI must augment rather than replace us or human workers are doomed

The article does not mention the workplace as the editorialized title would imply. It's primarily about trade unions.


I mean if they were targeting "software engineers" in general then Windows would be the obvious choice in 2026 as much as in 2006. But these early releases are all about the SF bubble where Mac is very much dominant.

Really? I frankly don’t know anyone who’s not on Linux. If you do any AI/ML you basically find yourself on a Linux box eventually. Perhaps I live in a bubble.

Surely it varies a lot and everyone is in an industry bubble to some extent, but from my experience in some non-tech industries (healthcare, manufacturing), Linux workstations were nonexistent and working with the Linux servers was more a sign of an ops role. People who wrote code for a living didn't touch them directly. Last StackOverflow survey [1] puts it at something like 50% use Windows at work, 30% Mac, 20-40% Linux (breakdown of WSL and Ubuntu as categories seems confusing, maybe the raw data is better).

[1] https://survey.stackoverflow.co/2025/technology/#1-computer-...


Yes, you live in a bubble.

I think a major factor in the hype is that it's especially useful to the kind of people with a megaphone: bloggers, freelance journalists, people with big social media accounts, youtubers, etc. A lot of project management and IFTTT-like automation type software gets discussed out of proportion to how niche it is for the same reason. Just something to keep in mind, I don't think it's some crypto conspiracy just a mismatch between the experiences of freelance writers vs everyone else.

While the popular thing when discussing the appeal of Clawdbot is to mention the lack of guardrails, personally I don't think that's very differentiating, every coding agent program has a command line flag to turn off the guardrails already and everyone knows that turning off the guardrails makes the agents extremely capable.

Based on using it lightly for a couple of days on a spare PC, the actual nice thing about Clawdbot is that every agent you create is automatically set up with a workspace containing plain text files for personalization, memories, a skills folder, and whatever folders you or the agents want to add. Everything being a plain text/markdown file makes managing multiple types of agents much more intuitive than other programs I've used which are mainly designed around having a "regular" agent which has all your configured system prompts and skills, and then hyperspecialized "task" agents which are meant to have a smaller system prompt, no persistent anything, and more JSON-heavy configuration. Your setup is easy to grok (in the original sense) and changing the model backend is just one command rather than porting everything to a different CLI tool.

Still, it does very much feel like using a vibe coded application and I suspect that for me, the advantages are going to be too small to put up with running a server that feels duct taped together. But I can definitely see the appeal for people who want to create tons of automations. It comes with a very good structure for multiple types of jobs (regular cron jobs, "heartbeat" jobs for delivering reminders and email summaries while having the context of your main assistant thread, and "lobster" jobs that have a framework for approval workflows), all with the capability to create and use persistent memories, and the flexibility to describe what you need and watch the agent build the perfect automation for it is something I don't think any similar local or cloud-based assistant can do without a lot of heavier customization.


They're offering 50% off the subscription to people who used to have Enhanced Autopilot [1]. As I predicted when the CEO's compensation plan had a part tied to FSD subscriptions, they are going to push more people onto it by bundling more features and cutting the price.

[1] https://x.com/SawyerMerritt/status/2014751111803032049


Show me the incentives i'll show you the outcomes.


Reminds me of when an ISP offered me a discount if I would agree to sign up with their partnered TV service. I agreed on the condition that I didn't have to rent a box. But you can't use the service without a box ... ? Who cares, I got a discount.


As hinted with the Finder comment, "Spotlight" is behind much more than the command-space search box. I don't know what the Siri services might do other than Siri itself, but wouldn't shock me if they were involved in things like Shortcuts and Control Center widgets. I understand thinking things you don't use are simply a "waste of CPU and storage space", but this reads like the kind of posts I used to see in the Windows XP era where people would open Task Manager and kill random processes they didn't understand. Best to make a little more effort to understand what the OS is doing before taking a scalpel to it. Or if you'd rather not, there's always OpenBSD (being serious here, it's pretty cool).


If some process is going to take hours of cpu time, it should be opt in. At a minimum I’d like to be able to turn the bloody things off if I don’t want them.

I run cpu usage meters in my menu bar. The efficiency cores always seem busy doing one thing or another on modern macOS. It feels like Apple treats my e-cores as a playground for stupid features that their developers want a lot more than I do - like photoanalysisd, or file indexing to power spotlight, that hasn’t worked how I want it to for a decade.

I have a Linux workstation, and the difference in responsiveness is incredible. Linux feels like a breath of fresh air. On a technical level, my workstation cpu is barely any faster. But it idles at 0%. Whenever I issue a command, I feel like the computer has been waiting for me and then it springs to action immediately.

To your point, I don’t care why these random processes are using all my cpu. I just want them to stop. I paid good money for my Apple laptop. The computer is for me. I didn’t pay all that money so some Apple engineer can vomit all over with their crappy, inefficient code.


Howard Oakley has made more effort to understand what macOS is doing (and blog about it) than practically anyone else ever.


I'm sure, this post just came off as confused.


What a trash article. Why is the only photo, used to illustrate the point about narrow buildings, a photo of Manhattan instead of anything in Japan? When "our zoning laws" are enumerated, where are they talking about? Last time I checked there were no US federal rules on parking spaces. At least they acknowledge that multiple jurisdictions exist when talking about health codes. And as per usual when talking about Japan, they ignore the fact that Japan also has car-dependent suburbs and rural areas, where it is quite common for restaurants outside of city centers to need to balance costs with the need for a larger footprint and a parking lot. The role of culture in eating habits is also ignored, Americans take more pride in the self-reliance of cooking their own meals.


Thanks for the questions. I used a picture of Manhattan intentionally to show that it is possible in some parts of America. There's already tons of pictures of that type of building in Japan, where it originated. The zoning laws are at the local and city level, as are the parking spaces. Japan does have car dependent suburbs and rural areas, I'm not saying they don't. It's likely that Japan's $4 meals are concentrated in not-rural areas. I really doubt that Americans take pride in not having cheap lunch options if they want them.


Seems like this article misses the enterprise angle which is the main question. I'm sure some gamers aching for an upgrade will sign up for cloud PCs while RAM is overpriced, just like how Geforce Now had a moment while GPUs were overpriced. But does it make any sense for businesses with massive fleets of Windows laptops, and might already have some kind of VDI setup, to replace them with thin clients? Would need some significant progress on the hardware.


I'm not really a big gamer but was looking into buying an xbox again. I already had a controller and thought why not try xbox cloud gaming on my Samsung TV.

With a decent internet connection I now struggle to see why anyone would want to buy a hardware Xbox. Games on the cloud version load instantly, play brilliantly and cost the same as the usual Game Pass as far as I can tell. The catalogue seems smaller maybe but aside from that I see little downside.

I could see it working well for PCs too - as long as the terminal device is seamless. I guess us devs have been renting computers in "the cloud" for decades anyway.


> I could see it working well for PCs too

I moonlight in film restoration. One 2hr movie out of our scanner is easily 16 TiB or more depending on the settings we scanned with.

Getting this uploaded to a remote server would take ~39hr over a fully-saturated 1Gbe pipe.


Clearly one use case where it wouldn't work.

On the other hand I'm a software engineer and my incredibly powerful MacBook could be not much more than a fancy dumb terminal - to be honest it almost is already.

If I can play a very responsive multiplayer game of the latest call of duty on my $300 TV with a little arm chip in it, then I could well imagine doing my job on a cloud Mac if the terminal device looked and felt like a MacBook but had the same tiny CPU my TV has.

Not sure if I'd choose it as a personal device but for corporations it seems a no brainer.


Sure, why not? A lot of them are already heavily invested in ms services. Where I work, laptops are on three year leases, they’d be easy to switch if the IT suits thought it would be cheap enough.


I'd say you are wrong on gamers aching for this. Any amount of latency ruins games, even turn based games lose a lot of their enjoyment when the ui starts getting delayed from user input.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: