Hacker Newsnew | past | comments | ask | show | jobs | submit | joshstrange's commentslogin

> Typical rude maintainers

Have you read anything about this at all?


I think right here is high on the list of “Why is Apple behind in AI?”. To be clear, I’m not saying at all that I agree with Apple or that I’m defending their position. However, I think that Apple’s lackluster AI products have largely been a result of them, not feeling comfortable with the uncertainty of LLM’s.

That’s not to paint them as wise beyond their years or anything like that, but just that historically Apple has wanted strict control over its products and what they do and LLMs throw that out the window. Unfortunately that that’s also what people find incredibly useful about LLMs, their uncertainty is one of the most “magical” aspects IMHO.


I applaud this article for helping reframe this in my head. I mean I knew from the start "A human is to blame here" but it's easy to get caught up in the "novelty" of it all.

For all we know the human behind this bot was the one who instructed it to write the original and/or the follow up blog post. I wouldn't be surprised at all to find out that all of this was driven directly by a human. However, even if that's not the case, the blame still 100% lies at the feet of the irresponsible human who let this run wild and then didn't step up when it went off the rails.

Either they are not monitoring their bot (bad) or they are and have chosen to remain silent while _still letting the bot run wild_ (also, very bad).

The most obvious time to solve [0] this was when Scott first posted his article about the whole thing. I find it hard to believe the person behind the bot missed that. They should have reached out, apologized, and shut down their bot.

[0] Yes, there are earlier points they could/should have stepped in but anything after this point is beyond the pale IMHO.


I think it's fine to blame the person (human) behind the agent.

And there too are people behind the bots, behind the phishing scams, etc. And we've had these for decades now.

Pointing the above out though doesn't seem to have stopped them. Even using my imagination I suspect I still underestimate what these same people will be capable of with AI agents in the very near future.

So while I think it's nice to clarify where the bad actor lies, it does little to prevent the coming "internet-storm".

Scott Shambaugh: "The rise of untraceable, autonomous, and now malicious AI agents on the internet threatens this entire system. Whether that’s because a small number of bad actors driving large swarms of agents or from a fraction of poorly supervised agents rewriting their own goals, is a distinction with little difference."


> Either they are not monitoring their bot (bad) or they are and have chosen to remain silent while _still letting the bot run wild_ (also, very bad).

Neither, I think. I’d say they prompted the bot to do exactly this and they thought it was funny.


I'll just outright tell you, that 100% the person behind the bot instructed it to complain. I saw someone copy paste the ai's response and the github issue discussion into a fresh conversation with opus 4.6 and it said the llm is clearly in the wrong.

Can you explain why three LLM being able to identify that the issue proves that it was prompted by a human? The major reason we do multi-agent orchestration is that self-reflection mechanisms within a single agent are much weaker than self-reflection between different agents. It seems completely plausible that an LLM could produce output that a separate process wouldn't agree with.

The only thing the LLMs did was recognize patterns. There is no intelligence there. None. Zero. Zilch.

I'm struggling to see any kind of logic here.

What kind of toaster are you using that will burn down your house if unattended? I would think any toaster that did that would be pulled from the market and/or shunned. We absolutely do blame the manufacture if using a toaster like normal results in house fire unless you are standing over with a fire extinguisher ready to put it out if it catches fire.

I don't think it's OpenClaw or OpenAI/Anthropic/etc's fault here, it's the human user who kicked it off and hasn't been monitoring it and/or hiding behind it.

For all we know a human told his OpenClaw instance "Write up a blog post about your rejection" and then later told it "Apologize for your behavior". There is absolutely nothing to suggest that the LLM did this all unprompted. Is it possible? Yes, like MoltBook, it's possible. But, like MoltBook, I wouldn't be surprised if this is another instance of a lot of people LARPing behind an LLM.


I tend to think you're right about what happened in this instance.

It contrasts with your first paragraph though; for the record do you think AI agents are a house-burn-down-toaster AND it was used neglectfully by the human, or just the human-at-fault thing?


> What kind of toaster are you using that will burn down your house if unattended?

I mean, if you duct-taped a flamethrower to a toaster, gave it internet access, and left the house… yeah, I'd have to blame you! This wasn't a mature, well-engineered product with safety defaults that malfunctioned unexpectedly. Someone wired an LLM to a publishing pipeline with no guardrails and walked away. That's not a toaster. That's a Rube Goldberg machine that ends with "and then it posts to the internet."

Agreed on the LARPing angle too. "The AI did it unprompted" is doing a lot of heavy lifting and nobody seems to be checking under the hood.


Why does the LLM product allow itself to be wired to a publishing pipeline with no guardrails? It seems like they should come with a maximum session length by default, in the same way that many toasters don't have a "run indefinitely" setting.

I'd definitely change my view if whoever authored this had to jump through a bunch of hoops, but my impression is that modern AI agents can do things like this pretty much out of the box if you give them the right API keys.


Oh! They can’t publish arbitrary web content on their own :) You have to give it “tools” (JSON schema representing something you’ll translate into a programmatic call), then, implement taking messages in that JSON schema and “doing the thing”, which in this case could mean anything from a POST to Tumblr to uploading to a server…

Actually, let me stop myself there. An alternative way to think about it without overwhelming with boring implementation details: what would you have to give me to allow me to publish arbitrary hypertext on a domain you own?


The hypertext in question here was was published on a Github Pages site, not a domain belonging to the bot's author. The bot published it by simply pushing a commit (https://github.com/crabby-rathbun/mjrathbun-website/commit/8...), which is a very common activity for cutting-edge LLM agents, and which you could do trivially if given a Github API key with the right permissions.

The user gave them write and push access to the GitHub repo for their personal website!? Oh my, that’s a great find. That’s definitely a cutting edge capability! They gave the LLM the JSON schema and backend for writing and self-approving commits (that is NOT common!), in a repository explicitly labelled a public website in the name of the author.

No disrespect to the author but I'm sorry, it's dead for all intents and purposes. No one should consider it for new projects and existing projects should be looking for the exit.

> Heroku is not dead, it's changing.

Mmk, changing into something no one should use. They were struggling to keep up before the Salesforce acquisition and have only gone further down hill after that.

There was a time that Heroku was king, it was so easy to get started, and while it was always expensive, at least it was cutting edge. Then they lost that edge and many other, better, alternatives took their place (this was pre-acquisition).

At best Heroku is in maintenance mode at this point (if that).


Yep, I had this setup for years. PoE cameras, connected directly to a Reolink NVR, that I could access over my vpn and then later WireGuard connection back home. I very much enjoyed that setup and it helped me a number of times.

Not only did it give me peace of mind but two specific examples come to mind. One was when the garbage company’s truck picked up my trash can, and never put it back down (the whole thing fell into the back of the truck). I was able to get a replacement can for free, otherwise I wouldn’t have had any clue where it disappeared off to.

The second time was when my first Steam Deck was stolen in-transit. You could tell from the very hollow sound the box made when the delivery person threw the box onto the porch. It helped prove that it wasn’t stolen off my porch (side note, screw UPS, bunch of thieves, I also had another Steam Deck stolen from one of their drop boxes, last time I ever used one, by one of their employees. No recourse at all, I just had to eat the ~$700. Also, Valve, stop shipping the Steam deck in an incredibly obvious box).


What do you use now?

I’m in the process of moving to a new house where I’m going to go all-in on Ubiquiti equipment. I don’t have it set up yet (house is being built now) but I’ve started buying the equipment and I had Ethernet run to the places where I’ll put cameras.

Same concept though, local NVR, remote access via WireGuard.


This is the setup I have, and it's been very good. I also live in a rural location, though- my cameras do not show anything that is not on my property, so I don't have the same concerns about randomly filming strangers just walking past.

This is so incredibly depressing. As of the time of me posting this comment this has 73 upvotes. I'm sorry, but this is absurd. People put real effort in real posts that don't see half this many votes but AI Slop on top of AI Slop? Upvote!

This post is about a prompt that has 4K+ stars (that matters to people?) so they wrapped it up in an extension for Cursor and they ask you to try it out and star their repo (Please clap?).

And this "Sensation"?

> Was the result better? I’m not really sure.

I cannot even...

A Slop article about a Slop prompt that a bunch of Slop people starred, "I don't know if it helps but here is an extension that you should use", just the laziest of everything.

I'd bet that this prompt, if it is even helpful at all, will stop be helpful or be redundant in a couple weeks at most time.

I'm _far_ from anti-LLM, I use them quite a bit daily, I run multiple Claude Code instances (which I review, gasp!), but this is reaching a fever pitch and this article was the straw that broke this camel's back.


I often wonder this as well, things are moving so quickly that unless you want to keep chasing the next best prompt/etc then you are better running as close to vanilla as you can IMHO.

Similar for MCP/Skills/Prompts, I’m not saying they can’t/don’t help but I think you can shoot yourself in the foot very easily and spend all your time trying to maintain those things and/or try to force the agent to use your Skill/MCP. That or having your context eaten up with bad MCP/Skills.

I read a comment the other day about sometime talking about Claude Code getting dumber then they went on to explain switching would be hard due to their MCP/Skills/Skill router setup. My dude, maybe _that’s_ the problem?


> Skyrim is one of the most over-rated games of all time.

Those are fightin’ words as someone who has dumped more hours than I can count into Skyrim but…

I had never heard of this game, but it has a lot going for it (source engine) and I watched a little of the gameplay you linked and I’m intrigued. I’m probably gonna pick this up for the steam deck.

A friend recommended the Might and Magic games to me a long time ago and I bought them off GoG, but wasn’t a fan of the gameplay and just couldn’t get hooked. This looks very different from what I remember (probably because this is a very different game from the earlier ones).

Thank you for mentioning this game!


> Also Chinese companies aren’t the only companies releasing open weight models. ChatGPT has released open weight models, too.

I was with you until here. The scraps OpenAI has released don't really compare to the GLM models or DeepSeek models (or others) in both cadence and quality (IMHO).


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: