codechicago277 21 hours ago

The fault lies entirely with the human operator for not understanding the risks of tying a model directly to the prod database, there’s no excuse for this, especially without backups.

To immediately turn around and try to bully the LLM the same way you would bully a human shows what kind of character this person has too. Of course the LLM is going to agree with you and accept blame, they’re literally trained to do that.

  • nominallyfree 8 hours ago

    I don't see the appeal of tooling that shields you from learning the admittedly annoying and largely accidental) complexity in developing software.

    It can only make accidental complexity grow and people's understanding diminish.

    When the inevitable problems become apparent, and you claim people should have understood better. Maybe using the tool that let's you avoid understanding things was a bad idea...

    • ben_w 6 hours ago

      Sure, but every abstraction does that.

      A manager hiring a team of real humans, vs. a manager hiring an AI, either way the manager doesn't know or learn how the system works.

      And asking doesn't help, you can ask both humans and AI, and they'll be different in their strengths and weaknesses in those answers, but they'll both have them — the humans' answers come with their own inferential distance and that can be hard to bridge.

clickety_clack a day ago

The whole thread seems very naive somehow. You can tell that he doesn’t fundamentally understand how a coding model works. The suggestion that it would know not to make any changes just because he said so means he doesn’t really understand what the model is. It’s built to generate (and apparently execute) code, so that is what it does. It doesn’t have an inner monologue running that says “ahh, a day off where I shoot the breeze around a whiteboard” or something. It’s more like an adderall addict with its fingers glued to the keyboard laying down all of its immediate thoughts directly as code with no forethought or strategy.

  • dimitri-vs 21 hours ago

    > I panicked and ran database commands without permission

    The AI responses are very suspicious. LLMs are extremely eager to please and I'm sure Replit system prompts them to err on the side of caution. I can't see what sequence of events could possibly lead any modern model to "accidentally" delete the entire DB.

    • maxbond 21 hours ago

      They're probabilistic. If it's possible, it'll happen eventually (and it is fundamental to language modeling that any sequence of tokens is possible). This is a straightforward Murphy's Law violation.

      • dimitri-vs 20 hours ago

        Maybe the individual tokens, but from experience of using LLMs something upstream encouraged the model to think it was okay to take the action of deleting the DB, something that would override safety RL, Replit system prompts and supposed user instructions not to do so. Just goes against the grain of every coding agent interaction I've ever had - seems fishy.

        • maxbond 19 hours ago

          According to the thread, the unit tests weren't passing, so the LLM reran the migration script, and the migration script blew out the tables. The "upstream encouragement" is a failing test.

          Is this a hoax for attention? It's possible, but the scenario is plausible, so I don't see reason to doubt it. Should I receive information indicating it's a hoax, I'll reassess.

    • Cthulhu_ 8 hours ago

      I think this debacle is actually a good learning opportunity for companies like this. If I were a decision maker in this space, I'd make it less magic or autonomous, and make it so that any critical operation is done by old fashioned boring but predictable programming, that is, "are you sure you want to drop database xyz?" dialogs.

Ecstatify a day ago

These AI-focused Twitter threads feel like they’re just recycling the same talking points for likes and retweets. When AI systems make mistakes, it doesn’t make sense to assign blame the way we would with human errors - they’re tools operating within their programming constraints, not autonomous agents making conscious choices.

  • mjr00 a day ago

    > When AI systems make mistakes, it doesn’t make sense to assign blame the way we would with human errors - they’re tools operating within their programming constraints, not autonomous agents making conscious choices.

    It's not really "assigning blame", it's more like "acknowledging limitations of the tools."

    Giving an LLM or "agent" access to your production servers or database is unwise, to say the least.

  • ayhanfuat a day ago

    I think at this point it is like rage-baiting. “AI wiped out my database”, “AI leaked my credentials”, “AI spent 2 million dollars on AWS” etc create interaction for these people.

    • phkahler a day ago

      The message reads like "AI did this bad thing" but we should all see it as "Another stupid person believed the AI hype and discovered it isn't trustworth" or whatever. You usually don't see them admit "gee that was dumb. What was I thinking?"

      • Cthulhu_ 8 hours ago

        Because that would mean they were wrong and their faith was misplaced. Faith is a good word to use in this case, because people like this are AI evangelists, going beyond selling it as "it is good because objective reasons 1, 2 and 3", into "this will revolutionize the world and how you think". They will overhype it and make excuses or talk around its flaws. Some of them are true believers, but I'm convinced most are just trying to sell a product or themselves.

  • blibble 21 hours ago

    the author is an ai booster

    he's not going to be happy with all this publicity

bluelightning2k 2 hours ago

Realistically - LLMs don't "delete the database". I find it quite unlikely that it proposed drop all out of nowhere. I wonder if what actually happened was a schema migration with an ORM? Prisma is still a pretty common choice, and FREQUENTLY migrations propose/require either a very nuanced path or a reset.

The second theory is an unbounded or inadequately bounded delete statement - essentially deleteMany on a single table.

From a more technical org I'd be interested in a write-up, but my intuition says one of those two paths to deleting technically a single table.

maxbond a day ago

Friends don't let friends run random untrusted code from the Internet. All code is presumed hostile until proven otherwise, even generated code. Giving an LLM write access to a production database is malpractice. On a long enough timeline, the likelihood of the LLM blowing up production approaches 1. This is the result you should expect.

  • maxbond 21 hours ago

    > Yesterday was biggest roller coaster yet. I got out of bed early, excited to get back @Replit ⠕ despite it constantly ignoring code freezes

    https://twitter-thread.com/t/1946239068691665187

    This wasn't even the first time "code freeze" had failed. The system did them the courtesy of groaning and creaking before collapsing.

    Develop an intuition about the systems you're building, don't outsource everything to AI. I've said before, unless it's the LLM who's responsible for the system and the LLM's reputation at stake, you should understand what you're deploying. An LLM with the potential to destroy your system violating a "code freeze" should cause you to change pants.

    Credit where it is do, they did ignore the LLM telling them recovery was impossible and did recover their database. And eventually (day 10), they did accept that "code freeze" wasn't a realistic expectation. Their eventual solution was to isolate the agent on a copy of the database that's safe to delete.

  • croes 2 hours ago

    Don't enter stranger's cars -> we got Uber

    Don't run foreign code from the Internet -> we got LLMs

Grimblewald a day ago

If you've ever tried getting a llm to solve moderatly difficult but solved tasks you'd know they're currently no good for anything beyond boilerplate code, and even then you have to watch it like a hawk.

consumer451 a day ago

I use LLM dev tools, and even have Supabase MCP running. I love these tools. They allowed me to create a SaaS product on my own, that I had no chance of creating otherwise as a long out of practice dev.

However, we are nowhere near the reliability of these tools to be able to:

1. Connect an MCP to a production database

2. Use database MCPs without a --read-only flag set, even on non-prod DBs

3. Doing any LLM based dev on prod/main. This obviously also applies to humans.

It's crazy to me that basic workflows like this are not enforced by all these LLM tools as they will save our mutual bacon. Are there any tools that do enforce using these concepts?

It feels like decision makers at these orgs are high on their own marketing, and are not putting necessary guardrails on their own tools.

Edit: Wait, even if we had AGI, wouldn't we still need things like feature branches and preview servers? Maybe the issue is that these are just crappy early tools missing a ton of features, and nothing to do with the reliability and power of LLMs?

  • Cthulhu_ 8 hours ago

    The only way LLM-based software development / production management will be trustable is by actually scaling back what it can and cannot do. Put critical operations in "real" code, so that the LLM can only request a release, triggering a human review of, at the very least, the operation that is about to be done.

    Then again, this reminds me of the prompts in operating systems whenever something needs root access, most people just blindly okayed it, especially on Windows since Vista did too many of them even for trivial operations.

  • avbanks 21 hours ago

    This imo is the biggest issue, LLMs can at times be very capable but they always are unreliable.

Arn_Thor 20 hours ago

This is the funniest thing I’ve seen in months. Maybe years? Incredible stuff

add-sub-mul-div a day ago

> I understand Replit is a tool, with flaws like every tool

> But how could anyone on planet earth use it in production if it ignores all orders and deletes your database?

Someday we'll figure out how to program computers deterministically. But, alas.

layer8 20 hours ago

One thing that AI likely won’t obviate the need of is making backups.

Here’s another funny one: https://aicodinghorrors.com/ai-went-straight-for-rm-rf-cmb5b...

  • rahimnathwani 20 hours ago

    Not only backups, but also a database with transaction logs or some way to play back the transactions after the most recent backup.

    • Cthulhu_ 8 hours ago

      TBH that's good practice without AIs too. People make mistakes, software has bugs, hardware dies.

mnafees 19 hours ago

One thing I’ve learned from seriously using AI agents for mundane coding tasks is: never ask them to do anything that involves deleting stuff. Incidents like these only reinforce that belief.

  • Cthulhu_ 8 hours ago

    If you use an agent they can do whatever they want IMO (caveat: I've never used one), but it's still your job to save and be able to revert the work (git) and to oversee anything involving production.

    It's like driving assistants, they feel like they can manage but in the end you are responsible.

cozzyd 15 hours ago

And here I was thinking AI had no sense of humor

nextaccountic a day ago

You need backups. If your lost data weren't due to AI slop, it could be a typo in a command, or anything else

blotfaba 15 hours ago

There was no database, it was a hoax.

cap11235 21 hours ago

> SaaStr.ai

Has to be a joke. Right?

  • swiftcoder 3 hours ago

    It's a real SaaS consultancy firm, at any rate

Alifatisk a day ago

Please do not link to twitter directly, use xcancel!

  • layer8 20 hours ago

    HN guidelines are to link to the original source, and Dang has confirmed that submissions shouldn’t link to mirror/proxy sites. Instead, circumventing links can be given in the comments.

krapht a day ago

Ahh, vibe coding.