this post was submitted on 11 Nov 2025
57 points (100.0% liked)

technology

24310 readers
237 users here now

On the road to fully automated luxury gay space communism.

Spreading Linux propaganda since 2020

Rules:

founded 5 years ago
MODERATORS
 

cross-posted from: https://lemmy.ml/post/38836048

you are viewing a single comment's thread
view the rest of the comments
[–] sodium_nitride@hexbear.net 20 points 4 months ago (1 children)

LLMs are good enough at coding now to replace the average fresh comp sci graduate.

This really isn't true. Modern LLMs are still not much better than an advanced Google search. I'm not even in CS doing industry work, but even I can spot misses in LLM output.

On a basic level, the AI has 2 major disadvantages. Firstly, it is not fully upto date with the Internet (and training on data past 2022 risks poisoning the dataset), a problem that will get worse over time. Secondly, LLMs are expensive as hell to actually run. Thirdly, the LLM context windows and higher order reasoning are still limited.

[–] gay_king_prince_charles@hexbear.net 3 points 4 months ago* (last edited 4 months ago) (1 children)

Firstly, it is not fully up to date with the Internet (and training on data past 2022 risks poisoning the dataset).

Where on earth did you get that from? Sonnet-4.5 has a pre-training cutoff date of January 2025 and GPT-5 has a pre-training cutoff date of October 2024. Any vaguely modern interface can get data past that into context by RAG and MCP. These aren't far back because of model collapse or anything, it's just that fine tuning is a hugely labor intensive process that takes months. Model collapse is greatly mitigated with human-based feedback and finetuning, making it safe to train models on LLM generated data. Deepseek, for example, is directly trained off GPT and Claude's output.

[–] sodium_nitride@hexbear.net 3 points 4 months ago

I am aware that LLMs do train with datasets past 2022. But there is a risk of poisoning the dataset that will grow overt time as the use of LLMs becomes larger. It is not a risk that can be easily mitigated by human feedback and fine tuning, since getting rid of workers is exactly why business owners are hyped about LLMs in the first place.

And yes, I did not about MCP so I was wrong about that part, but you can still put less data into context vs in training.