this post was submitted on 07 Mar 2026
840 points (97.3% liked)

Technology

82460 readers
4336 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] artyom@piefed.social -5 points 2 days ago (1 children)

It can, but we've already seen many times that it does not.

[–] suicidaleggroll@lemmy.world 4 points 2 days ago* (last edited 2 days ago) (1 children)

Only if the user has configured it to bypass those authorizations.

With an agentic coding assistant, the LLM does not decide when it does and doesn’t prompt for authorization to proceed. The surrounding software is the one that makes that call, which is a normal program with hard guardrails in place. The only way to bypass the authorization prompts is to configure that software to bypass them. Many do allow that option, but of course you should only do so when operating in a sandbox.

The person in this article was a moron, that’s all there is to it. They ran the LLM on their live system, with no sandbox, went out of their way to remove all guardrails, and had no backup. The fallout is 100% on them.

[–] artyom@piefed.social -3 points 2 days ago (1 children)

As I said elsewhere, if you're denying access to your agentic AI, what is the point of it? It needs access to complete agentic tasks.

The person in this article was a moron, that’s all there is to it. They ran the LLM

No disagreement there.

[–] suicidaleggroll@lemmy.world 2 points 2 days ago* (last edited 2 days ago) (1 children)

if you’re denying access to your agentic AI, what is the point of it? It needs access to complete agentic tasks.

Yes, which it can prompt you for. Three options:

  1. Deny everything
  2. Prompt for approval when it needs to run a command or write a file
  3. Allow everything

Obviously optional 1 is useless, but there’s nothing wrong with choosing option 2, or even option 3 if you run it in a sandbox where it can’t do any real-world damage.

[–] artyom@piefed.social -3 points 2 days ago (1 children)
  1. Prompt for approval when it needs to run a command or write a file

And therein lies the problem. You're giving the LLM control over when to or not to ask for approval.

[–] suicidaleggroll@lemmy.world 2 points 2 days ago

You clearly have absolutely zero experience here. When you're prompted for access, it tells you the exact command that's going to be run. You don't just give blind approval to "run something", you're shown the exact command it's going to run and you can choose to approve or reject it.