this post was submitted on 04 Apr 2026
70 points (98.6% liked)
technology
24315 readers
356 users here now
On the road to fully automated luxury gay space communism.
Spreading Linux propaganda since 2020
- Ways to run Microsoft/Adobe and more on Linux
- The Ultimate FOSS Guide For Android
- Great libre software on Windows
- Hey you, the lib still using Chrome. Read this post!
Rules:
- 1. Obviously abide by the sitewide code of conduct. Bigotry will be met with an immediate ban
- 2. This community is about technology. Offtopic is permitted as long as it is kept in the comment sections
- 3. Although this is not /c/libre, FOSS related posting is tolerated, and even welcome in the case of effort posts
- 4. We believe technology should be liberating. As such, avoid promoting proprietary and/or bourgeois technology
- 5. Explanatory posts to correct the potential mistakes a comrade made in a post of their own are allowed, as long as they remain respectful
- 6. No crypto (Bitcoin, NFT, etc.) speculation, unless it is purely informative and not too cringe
- 7. Absolutely no tech bro shit. If you have a good opinion of Silicon Valley billionaires please manifest yourself so we can ban you.
founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments

The free models are much worse than the $500 per user/month enterprise ones. I have seen these be able to generate working features first hand at work, and I cannot deny that certain models are capable of implementing features when appropriate requirements are provided. To claim anything else would be to deny what I have seen with my own eyes.
However, therein lies the trap. Just because it is capable of achieving the provided task in one instance, doesn't mean that it always provides an appropriate answer or solution in all cases.
But those who have initially used it successfully tend to start believing its output uncritically. I've noticed this on myself when I tried it at work, and I think this is basic human, heck, even animal condition. You are naturally inclined to trust an entity that initially provides you with beneficial output. You become less critical, as the output often sounds informed and convincing, and in many cases provably works as well (especially when a robust testing framework exists inside the project. its only through unit and integration tests that these AIs can even reliably implement features).
But this leads to an increasing reliance on the tech, and you stop being capable of arguing why the solution it generated works. You have to put in active effort to question what it's doing, and you have no way of knowing whether it's telling you the truth or lies, because it has no motive, and researching the facts can take so long that it completely defeats the point of automation. So it ends up being rather self-defeating in many cases, and can leave you less capable of solving problems yourself.
I think the most useful application for it personally is to use it for debugging -- feed it a cryptic error message, and it will usually generate an answer that, while not necessarily accurate, can give you more pointers to find the true answer, much better than most search engines can.
I mean deepseek will make you working programs for 20 cents of tokens sometimes if the requirements are straightforward and it's nothing too exotic.