168
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
this post was submitted on 22 Jul 2023
168 points (85.6% liked)
Asklemmy
43750 readers
1250 users here now
A loosely moderated place to ask open-ended questions
If your post meets the following criteria, it's welcome here!
- Open-ended question
- Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
- Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
- Not ad nauseam inducing: please make sure it is a question that would be new to most members
- An actual topic of discussion
Looking for support?
Looking for a community?
- Lemmyverse: community search
- sub.rehab: maps old subreddits to fediverse options, marks official as such
- !lemmy411@lemmy.ca: a community for finding communities
~Icon~ ~by~ ~@Double_A@discuss.tchncs.de~
founded 5 years ago
MODERATORS
It's a language model, it can't even do math reliably. Yes, it produces code that works sometimes, but it also hallucinates functions that don't exist or can introduce bugs you won't notice at first glance.
And writing a script is different than extending an existing code base. How often do you really start a greenfield project?
I wouldn't even know how to input a code base into ChatGPT to extend, do you just throw in hundreds of files with a 100k+ lines of code?
I guess LLM with plugins can solve most of the problems. ChatGPT can already interact with Wolfram Alpha to do math.
I can imagine similar plugins for code. Like it knows what kind of function it needs, so it interacts with a plugin that searches the code base to see if it exists. It might get back a snippets of candidates and examples how they’re used in the code already.
This is probably a difficult thing to achieve, but I don’t think it’s impossible. It’s probably going to take some time until something like this is made.