1332
submitted 2 months ago by floofloof@lemmy.ca to c/technology@lemmy.world
you are viewing a single comment's thread
view the rest of the comments
[-] N0body@lemmy.dbzer0.com 316 points 2 months ago

There’s an alternate timeline where the non-profit side of the company won, Altman the Conman was booted and exposed, and OpenAI kept developing machine learning in a way that actually benefits actual use cases.

Cancer screenings approved by a doctor could be accurate enough to save so many lives and so much suffering through early detection.

Instead, Altman turned a promising technology into a meme stock with a product released too early to ever fix properly.

[-] patatahooligan@lemmy.world 65 points 2 months ago

No, there isn't really any such alternate timeline. Good honest causes are not profitable enough to survive against the startup scams. Even if the non-profit side won internally, OpenAI would just be left behind, funding would go to its competitors, and OpenAI would shut down. Unless you mean a radically different alternate timeline where our economic system is fundamentally different.

[-] rsuri@lemmy.world 57 points 2 months ago* (last edited 2 months ago)

I mean wikipedia managed to do it. It just requires honest people to retain control long enough. I think it was allowed to happen in wikipedia's case because the wealthiest/greediest people hadn't caught on to the potential yet.

There's probably an alternate timeline where wikipedia is a social network with paid verification by corporate interests who write articles about their own companies and state-funded accounts spreading conspiracy theories.

[-] Petter1@lemm.ee 12 points 2 months ago

There are infinite timelines, so, it has to exist some(wehere/when/[insert w word for additional dimension]).

[-] msage@programming.dev 8 points 2 months ago

What is OpenAI doing with cancer screening?

[-] mustbe3to20signs@feddit.org 29 points 2 months ago

AI models can outmatch most oncologists and radiologists in recognition of early tumor stages in MRI and CT scans.
Further developing this strength could lead to earlier diagnosis with less-invasive methods saving not only countless live and prolonging the remaining quality life time for the individual but also save a shit ton of money.

[-] T156@lemmy.world 33 points 2 months ago

That is a different kind of machine learning model, though.

You can't just plug in your pathology images into their multimodal generative models, and expect it to pop out something usable.

And those image recognition models aren't something OpenAI is currently working on, iirc.

[-] mustbe3to20signs@feddit.org 16 points 2 months ago

I'm fully aware that those are different machine learning models but instead of focussing on LLMs with only limited use for mankind, advancing on Image Recognition models would have been much better.

[-] Grandwolf319@sh.itjust.works 6 points 2 months ago

I agree but I also like to point out that the AI craze started with LLMs and those MLs have been around before OpenAI.

So if openAI never released chat GPT, it wouldn’t have become synonymous with crypto in terms of false promises.

[-] tfowinder@lemmy.ml 3 points 2 months ago

Don't know about image recognition but they released DALL-E , which is image generating and in painting model.

[-] Grandwolf319@sh.itjust.works 3 points 2 months ago

Not only that, image analysis and statistical guesses have always been around and do not need ML to work. It’s just one more tool in the toolbox.

[-] Petter1@lemm.ee 2 points 2 months ago* (last edited 2 months ago)

Fun thing is, most of the things AI can, they never planned it to be able to do it. All they tried to achieve was auto completion tool.

[-] msage@programming.dev 14 points 2 months ago

Wasn't it proven that AI was having amazing results, because it noticed the cancer screens had doctors signature at the bottom? Or did they make another run with signatures hidden?

[-] mustbe3to20signs@feddit.org 10 points 2 months ago* (last edited 2 months ago)

There were more than one system proven to "cheat" through biased training materials. One model used to tell duck and chicken apart because it was trained with pictures of ducks in the water and chicken on a sandy ground, if I remember correctly.
Since multiple medical image recognition systems are in development, I can't imagine they're all ~~this faulty~~ trained with unsuitable materials.

[-] msage@programming.dev 6 points 2 months ago

They are not 'faulty', they have been fed wrong training data.

This is the most important aspect of any AI - it's only as good as the training dataset is. If you don't know the dataset, you know nothing about the AI.

That's why every claim of 'super efficient AI' need to be investigated deeper. But that goes against line-goes-up principle. So don't expect that to happen a lot.

[-] Petter1@lemm.ee 2 points 2 months ago

Or we get to a time where we send a reprogrammed terminator back in time to kill altman 🤓

this post was submitted on 27 Sep 2024
1332 points (99.4% liked)

Technology

60033 readers
2914 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS