1242
you are viewing a single comment's thread
view the rest of the comments
[-] Hazzard@lemm.ee 16 points 7 months ago

I don't necessarily disagree that we may figure out AGI, and even that LLM research may help us get there, but frankly, I don't think an LLM will actually be any part of an AGI system.

Because fundamentally it doesn't understand the words it's writing. The more I play with and learn about it, the more it feels like a glorified autocomplete/autocorrect. I suspect issues like hallucination and "Waluigis" or "jailbreaks" are fundamental issues for a language model trying to complete a story, compared to an actual intelligence with a purpose.

this post was submitted on 10 Apr 2024
1242 points (99.0% liked)

Programmer Humor

19572 readers
978 users here now

Welcome to Programmer Humor!

This is a place where you can post jokes, memes, humor, etc. related to programming!

For sharing awful code theres also Programming Horror.

Rules

founded 1 year ago
MODERATORS