210
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
this post was submitted on 26 Mar 2024
210 points (91.7% liked)
Technology
59456 readers
4201 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
That is literally a complete misinterpretation of how models work.
You don't "have the Internet as a model", you train a model using large amounts of data. That does not mean, that this model contains any of the actual data. State of the at models are somewhere in the billions of parameters. If you have, say, 50b parameters, each being a 64bit/8 byte double (which is way, way too much accuracy) you get something like 400gb of data. That's a lot, but the Internet slightly larger than that.
It's an exaggeration, but its not far off given that Google literally has all of the web parsed at least once a day.
Reddit just sold off AI harvesting rights on all of its content to Google.
The problem is no longer model size. The problem is interpretation.
You can ask almost everyone on earth a simple deterministic math problem and you'll get the right answer almost all of the time because they understand the principles behind it.
Until you can show deterministic understanding in AI, you have a glorified chat bot.
It is far off. It's like saying you have the entire knowledge of all physics because you skimmed a textbook once.
Interpretation is also a problem that can be solved, current models do understand quite a lot of nuance, subtext and implicit context.
But you're moving the goal post here. We started at "don't get better, at a plateau" and now you're aiming for perfection.
You're building beautiful straw men. They're lies, but great job.
I said originally that we need to improve the interpretation of the model by AI, not just have even bigger models that will invariably have the same flaw as they do now.
Deterministic reliability is the end goal of that.
Where exactly did you write anything about interpretation? Getting "details right" by processing faster? I would hardly call that "interpretation" that's just being wrong faster.