People Twitter
People tweeting stuff. We allow tweets from anyone.
RULES:
- Mark NSFW content.
- No doxxing people.
- Must be a pic of the tweet or similar. No direct links to the tweet.
- No bullying or international politcs
- Be excellent to each other.
- Provide an archived link to the tweet (or similar) being shown if it's a major figure or a politician.
it only takes a couple times of getting a made-up bullshit answer from chatgpt to learn your lesson of just skip asking chatgpt anything altogether
That's what people get when they ask me questions too but they still bother me all the time so clearly that's not going to work.
I stopped using it when I asked who I was and then it said I was a prolific author then proceeded to name various books I absolutely did not write.
I just read "The Autobiography of QueenHawlSera"!
Have I been duped?
Why the fuck would it know who you are?
If you have an account, you can tell it things about yourself. I used my boss's account for a project at work (felt gross). I made the mistake of saying "good morning" to it one day, and it proceeded to ask me if I was going to do (activities related to my boss's personal life - and the details were accurate). I was thinking, "why does he tell it so much about himself?"
and I'm apparently a famous Tiktoker and Youtuber.
I was using it to blow through an online math course I'd ultimately decided I didn't need but didn't want to drop. One step of a problem I had it solve involved finding the square root of something; it spat out a number that was kind of close, but functionally unusable. I told it it made a mistake three times and it gave a different number each time. When I finally gave it the right answer and asked, "are you running a calculation or just making up a number" it said that if I logged in, it would use real time calculations. Logged in on a different device, asked the same question, it again made up a number, but when I pointed it out, it corrected itself on the first try. Very janky.
ChatGPT doesn't actually do calculations. It can generate code that will actually calculate the answer, or provide a formula, but ChatGPT cannot do math.
You need multi-shot prompting when it comes to math. Either the motherfucker gets it right, or you will not be able to course correct it in a lot of cases. When a token is in the context, it's in the context and you're fucked.
Alternatively you could edit the context, correct the parameters and then run it again.
On the other side of the shit aisle
Shoutout to my man Mistral Small 24B who is so insecure, it will talk itself out of correct answers. It's so much like me in not having any self worth or confidence.
I've only really found it useful when you provide the source of information/data to your prompt. E.g. say you want to convert one data format to another like table data into JSON
It works very consistently in those types of use cases. Otherwise it's a dice roll.
First they came for my mansplaining and I said nothing...
If ai can start pirating old movies then it's curtains for me brotherzz 😞
"Hey Alexa, add Rick and Morty to Sonarr and download the first season for me"
You got a girlfriend with that skill? You have no idea how much hope you've given me my brother.
For real, I feel like its part of the reason Ive kept my S/O so long 😭
People do know they could have just googled things and looked at the top results before right? Even with all the enshitification it's gone through it will still usually yield something useful for any topic that isn't super niche. Like if I search "shorting" the top 3 results are articles from Wikipedia, Investorpedia, (don't know enough about it to know if it's reliable) and Charles Schwab, a source with a conflict of interest since they probably sell that service, but they're probably at least going to explain why you'd want to buy it from them, and then as a bonus the 4th result is an ELI5 Reddit thread, which while probably not the most reliable source of info, is probably about on the same level as randomly asking your SO about a topic which they're not an expert in.
Yeah but people now Google things and look at the AI summary of the top 3 results.
Just fyi investorpedia is reliable. Ive personally never found incorrect information there.
Im not so sure id call schwab a conflict of interest exactly. They do provide the service, but financial institutions like that are not allowed to provide investing advice, so they cant upsell strategies like that. All they can do is explain it so that you know what strategy you want to pursue.