624
submitted 18 hours ago by misk@sopuli.xyz to c/technology@lemmy.world
top 50 comments
sorted by: hot top controversial new old
[-] sirico@feddit.uk 4 points 41 minutes ago

Always invest in the spades never the gold mine

[-] fluxion@lemmy.world 4 points 55 minutes ago

AI companies specializing in spreading bullshit all across the internet have a bright future however

[-] TehWorld@lemmy.world 4 points 1 hour ago

So, I have clients that are actively using AI on a daily basis and LOVE it. It is however a very narrow subset. Also, I’m pretty sure that a LARGE amount of Dollars are currently being spent on AI generated political articles.

[-] bluewing@lemm.ee 26 points 3 hours ago

No shit.

Like all new technologies, there is a time when bunches of companies jump on the band wagon to get in on the action. You can see it all throughout the history of the industrial revolution.

They mostly know that there will come a great weeding out of those that can't handle the technology or just fail from poor management. But they are betting they will be among the 1% that wins the race and remain to dominate the market.

The rest will just bide their time until the next Big Thing comes along. And the process starts over again.

[-] Blackmist@feddit.uk 33 points 5 hours ago

Yeah, but thanks to the glory of corporateworld, all the people involved in making these decisions will be in a higher position at a different company by the time the consequences come knocking.

You definitely will not regret spending billions of dollars on GPUs and electricity bills.

[-] GeneralInterest@lemmy.world 29 points 6 hours ago

Maybe it's like the dotcom bubble: there is genuinely useful tech that has recently emerged, but too many companies are trying to jump on the bandwagon.

LLMs do seem genuinely useful to me, but of course they have limitations.

[-] linearchaos@lemmy.world 7 points 3 hours ago

We need to stop viewing it as artificial intelligence. The parts that are worth money are just more advanced versions of machine learning.

Being able to assimilate a few dozen textbooks and pass a bar exam is a neat parlor trick, but it is still just a parlor trick.

Unfortunately probably the biggest thing to come out of it will be the marketing aspect. If they spend enough money to train small models on our wants and likes it will give them tremendous amounts of return.

The key to using it in a financially successful manner is finding problems that fit the bill. Training costs are fairly high, quality content generation is also rather expensive. There are sticky problems around training it from non-free data. Whatever you're going to use it for either needs to have a significant enough advantage to make the cost of training /data worth it.

I still think we're eventually going to see education rise. The existing tools for small content generation adobe's use of it to fill in small areas is leaps and bounds better than the old content aware patches. We've been using it for ages for speech recognition and speech generation. From there it's relatively good at helper roles. Minor application development, copy editing, maybe some VFX generation eventually. Things where you still need a talented individual to oversee it but it can help lessen the workload.

There are lots of places where it's being used where I think it's a particularly poor fit. AI help desk chatbots, IVR scenarios, It says brain dead as the original phone trees and flow charts that we've been following for decades.

[-] Eheran@lemmy.world -2 points 1 hour ago

If GPT4o is still not what you would call AI, then what is? You can have conversations with it, the Turing test is completely irrelevant all of the sudden.

[-] Buddahriffic@lemmy.world 2 points 55 minutes ago

It's a massive text predictor. It doesn't solve problems, it applies patterns based on correlations it picked up during training. If someone talked about your topic online, it has been trained on those conversations. If a topic has two sides that don't agree, chat gpt might respond in a way that is biased towards one side or the other and you can easily get it to "switch" to the other side with follow up prompts.

For what would be considered AI, think of the star trek computer or Data. The Star Trek computer could create simulations of warp core behaviour to push frontiers of knowledge or characters smart enough to defeat its own safeties (frankly, the computer was such a deus ex machina kinda thing that it was hard to suspend disbelief at times, like why did they even have humans doing the problem solving with computers that capable?). Data wouldn't get confused about whether any counties in Africa start with K.

I don't think the Turing test is an effective means of determining intelligence anyways. It came from a time when a conversational computer was barely thinkable. But I wouldn't even say chat gpt is there yet, since you can tell if you ask it the right things. It is very useful, don't get me wrong, like a very powerful search engine. But it's not intelligent.

[-] Eheran@lemmy.world 1 points 51 minutes ago

What of what you say does not apply to humans? They apply patterns of behavior in response to some input. Picked up by learning them. Including people talking online. They are always biased on some way. Some will acknowledge their bias and change it if you give them context.

GPT can literally create simulations. I have used it to do exactly that, specifically for 2D heat conducting with coupled mass transport and reaction kinetics.

[-] Buddahriffic@lemmy.world 1 points 5 minutes ago

Yeah, it does do some very human-like things, but it's still missing some important parts.

It's kinda like using a textbook for problem solving. It's great at helping you solve instances of problems that have already been solved, but you won't likely find the next big advancement in that field in a textbook.

Newton realized masses attracted each other, and through experimentation, came up with his laws of classical physics.

Einstein took the idea that the speed of light always seems to be the same despite relative motion to come up with special relativity, then realized that space-time itself was a physical thing that could be interacted with rather than just a medium, plus came up with field equations that were used to predict things like black holes before anyone had any kind of notion that they were real things.

Chat gpt is incapable of things like that. And sure, many humans never do anything like that, some might not even be capable even if they were motivated and had the right supports to try. But many humans do solve problems that they've never seen before. There's big names in academia but so many more that don't get famous but still push the boundaries of human knowledge, creatively solving problems and answering questions every day.

I wouldn't be surprised if an LLM is a piece of general AI if or when it comes, but there will be other parts that are currently missing. We don't even know what consciousness is, let alone if any of our hardware is capable of creating/hosting one.

[-] Cryophilia@lemmy.world 0 points 1 hour ago

I can write a program that just replies "yes" to everything you say and you can have a conversation with that. Is that program AI?

"AI isn't really AI and no one ever thought that AI was actually AI so it doesn't matter if we call it AI" is the funniest level of tech bro cope these days.

[-] Womble@lemmy.world 2 points 35 minutes ago

AI has been the name of the field for 70 years at this point, it isn't something Sam Altman came up with as a marketing wheeze.

[-] TotalFat@lemmy.world 1 points 23 minutes ago

Magic Eightball

[-] datelmd5sum@lemmy.world 7 points 5 hours ago

We're hitting logarithmic scaling with the model trainings. GPT-5 is going to cost 10x more than GPT-4 to train, but are people going to pay $200 / month for the gpt-5 subscription?

[-] Skates@feddit.nl 3 points 2 hours ago

Is it necessary to pay more, or is it enough to just pay for more time? If the product is good, it will be used.

[-] madis@lemm.ee 3 points 4 hours ago

But it would use less energy afterwards? At least that was claimed with the 4o model for example.

4o is also not really much better than 4, they likely just optimized it among others by reducing the model size. IME the "intelligence" has somewhat degraded over time. Also bigger Model (which in tha past was the deciding factor for better intelligence) needs more energy, and GPT5 will likely be much bigger than 4 unless they somehow make a breakthrough with the training/optimization of the model...

[-] hglman@lemmy.ml 2 points 3 hours ago

4o is optimization of the model evaluation phase. The loss of intelligence is due to the addition of more and more safeguards and constraints by the use of adjunct models doing fine turning, or just rules that limit whole classes of responses.

load more comments (1 replies)
[-] bamfic@lemmy.world 53 points 9 hours ago* (last edited 9 hours ago)

I am old enough to remember when the CEO of Nortel Networks got crucified by Wall Street for saying in a press conference that the telecom/internet/carrier boom was a bubble, and the fundamentals weren't there (who is going to pay for long distance anymore when calls are free over the internet? where are the carriers-- Nortel's customers-- going to get their income from?). And 4 years later Nortel ceased to exist. Cisco crashed too, though had enough TCP/IP router biz and enterprise sales to keep them alive even until today.

This all reminds me of the late 1990s internet bubble rather than the more recent crypto bubble. We'll all still be using ML models for all kinds of things more or less forever from now on, but it won't be this idiotic hype cycle and overvaluation anymore after the crash.

Shit, crypto isn't going anywhere either, it's a permanent fixture now, Wall Street bought into it and you can buy crypto ETFs from your stockbroker. We just don't have to listen to hype about it anymore.

Crypto is still just as awful as it ever was IMO. Still plenty of assholes ~~gambling~~ investing in crypto.

[-] wrekone@lemmyf.uk 12 points 7 hours ago

Well put.

Soon, it won't be this idiotic hype cycle, but it'll be some other idiotic hype cycle. Short term investors love hype cycles.

[-] kautau@lemmy.world 7 points 9 hours ago

We just don’t have to listen to the hype about it anymore.

True, it’s now in most circles just been mixed in as a commodity to trade on. Though I wish everyone would get that. There’s still plenty of idiots with .eth usernames who think there’s some new boon to be made. The only “apps” built on crypto networks were and are purely for trading crypto, I’ve never seen any real tangible benefit to society come out of it. It’s still used plenty for money laundering, but regulators are (slowly) catching up. And it’s still by far the easiest way to demonstrate what happens to unregulated markets.

https://www.web3isgoinggreat.com/

load more comments (12 replies)
[-] ulkesh@lemmy.world 6 points 8 hours ago* (last edited 8 hours ago)

Wow, a CEO who doesn’t buy into the hype? That’s astonishing.

I, for one, cannot wait for the bubble to burst so we can get back to some sense of sanity.

Edit>> Though if Baidu is investing in AI like all the rest, then maybe they just think they’ll be immune — in which case I’m sad again that I haven’t yet come across a CEO who calls bullshit on this nonsense.

AI will have its uses, and it has practical use cases such as helping people to walk or to speak or to translate in real time, etc. But we’re decades away from what all these CEOs seem to think they’re going to cash in on now. And it’ll be fun on some level watching them all be wrong.

[-] nyan@lemmy.cafe 1 points 2 hours ago

Edit>> Though if Baidu is investing in AI like all the rest, then maybe they just think they’ll be immune — in which case I’m sad again that I haven’t yet come across a CEO who calls bullshit on this nonsense.

They may just have kept their AI investments responsible—that is, not put more money into it than they can afford to lose. Keep in mind, Baidu is the Chinese equivalent of Google. They have a large, diversified business with many income streams. I expect they'll still be around after the bubble bursts.

load more comments
view more: next ›
this post was submitted on 21 Oct 2024
624 points (97.7% liked)

Technology

58810 readers
3885 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS