GamingChairModel

joined 2 years ago
[–] GamingChairModel@lemmy.world 1 points 3 hours ago

Don't they put plutonium reactors in space?

The ones that power spacecraft generate less than 5000W of heat at max power (while producing 300W of usable electricity).

In order to power a single server rack of 72 Blackwell GPUs, which takes about 130,000 watts, you'd need about 430 of those RTGs, and need to manage cooling requirements of 430 times as much (plus however much additional power will be required by the cooling system itself, too).

[–] GamingChairModel@lemmy.world 8 points 3 hours ago

Companies are building entire workflows around AI, but they are building them under the assumption that they won't ever be charged per token.

Or worse, where the AI models underpinning a workflow breaks or degrades in some way to reduce token usage and then starts behaving in unexpected ways, in a process/workflow that assumes a particular type of behavior.

Counterpoint: sometimes the best still shot requires a particular moment captured with a particular, consciously arranged setup.

This interview of a veteran NBA photographer breaks it down of how he only has a single shot per shot because of how he necessarily relies on strobes set up to not distract the players or interfere with the broadcast. As a result, he scouts/studies each player and team so that he knows when the right moment is to actually capture the shot, because he can't exactly ask players to do it again.

If you read interviews of Pulitzer photography winners, they'll often say a lot of the same things: being prepared and being lucky and having that convergence of having incredibly high skill/expertise/understanding of the setting, while being able to capture in every opportunity presented.

You should capture a lot of photos and examine them to understand how to make them better, and increase your skill level and understand your subject so that you can still optimize for the very best shot possible.

Original reporting by Bloomberg is here or here for an archive.is version.

Sounds like the talks are stalled about revenue guarantees from the Kenyan government, if demand for the data center's capacity never shows up. The power infrastructure isn't really an issue yet, since the first phase is going to be 100MW and there are plans to build geothermal plants sufficient to cover several multiples of the country's current power usage, including whatever demand comes from this data center.

I used to carry around an 11" laptop when I had to give a bunch of presentations in unfamiliar conference rooms. It was the main reason why I switched to Apple around 15 years ago for my travel laptop (couldn't get Linux-friendly hardware with good battery life and seamless display/audio support over DVI/DP/HDMI and whatever audio setup they'd have for me) while still keeping a "main" Linux laptop for around the house and a headless server sitting next to my router.

I'd love to have that 11" form factor again, with modern thunderbolt/USB4 docking stations at my work desk, home office, and whatever desk I might hotel at or whatever. I rarely used the small screen or keyboard but it was nice to have that option on the move (like in an economy plane seat).

Five star rating system was dumb because almost every rating was 1 or 5 stars. It was right to replace with a thumbs up/thumbs down system.

That assumes that the only use for ratings is for averaging the aggregate votes across all users. Nope. Sometimes for a specific user they like to be able to see the granularity of their own ratings for their own use. And even if it is a public aggregated thing the rating service can still treat all 1-2 stars as downvotes and 4-5 stars as upvotes while it's easier to use the simpler algorithms, but to still store the more precise data for analyzing correlations at greater detail.

Big tech covered the world in trillion-parameter AI models and couldn't even figure out what to do with 5-star ratings differently from upvotes/downvotes? It's ridiculous.

[–] GamingChairModel@lemmy.world 5 points 2 days ago (3 children)

Yeah if I were starting now I'd be looking at jellyfin. But I paid the lifetime plex pass, and inertia/laziness what it is, so I haven't found a reason to actually switch yet.

Transformers are like blockchain: an interesting use of mathematical principles to solve certain problems in a novel way, where the hype around that core attracts charlatans and scammers and combinations of the two traits who claim that it will go on to solve totally different problems in such a way as to revolutionize the world we live in.

NFTs were the end of that line for blockchain where the machine started to eat itself. I can see a future, stable use of blockchain in some limited contexts, but cryptobros have always overstated the contexts in which that particular type of digital ledger can be more useful than other types of digital ledgers.

We'll see where the end of the road is for transformers, and what's left at the end. I believe that computer inference will always be useful in some contexts, and that the advances in huge models with absurdly large numbers of parameters have unlocked some previously impractical tasks, but I could also see that settling into a general background existence as just another technological tool for doing things in a world that still looks pretty similar to the world today.

Hey now, that infrastructure is good for, like, 3 years, so it's really like spending $145 billion to save $81 billion. And that doesn't even start to get into how much it costs to operate that infrastructure.

[–] GamingChairModel@lemmy.world 1 points 2 days ago (1 children)

Their stubbornness on the mouse front (including refusal to add a second mouse button) at least got their trackpads (and associated gestures and haptics) to be the best in the game.

Still, I prefer using a mouse if I'm actually at a desk, so I just hook up a Logitech to my MacBook.

[–] GamingChairModel@lemmy.world 11 points 4 days ago

I imagine a lot of the companies will realize that the purported benefits of AI have hidden defects they have to fix, piling up lots of hidden technical/administrative debt, right around the time their AI vendors jack up prices or go out of business so that they're stuck with unmaintainable stuff (code, workflows, processes) that they don't fully understand, built by and for AI agents that no longer exist.

[–] GamingChairModel@lemmy.world 2 points 5 days ago (1 children)

And the other way, too. The whole reason why Android chose Java was because it was, at the time, one of the better languages and runtimes for creating hardware-agnostic software. Now that a software ecosystem is in place, why should Google be able to control what hardware the already-written software runs on?

 

I've read some of Ed Zitron's long posts on why the AI industry is a bubble that will never be profitable (and will bring down a lot of companies and investors), and one of the recurring themes is that the AI companies are trying to capture growing market share in an industry where their marginal profits are still negative, and that any increase in revenue necessarily increases their costs of providing their services.

But some of the comments in various HackerNews threads are dismissive, saying that each new generation of models makes the cost of inference lower, so that with sufficient customer volume, the companies running the models can make enough profit on inference to make up for the staggering up-front capital expenditures it took to build out the data centers, train their models, etc.

It's all pretty confusing to me. So for those of you who are familiar with the industry, I have several questions:

  1. Is the cost of running any given pretrained model going down, for specific models? Are there hardware and software improvements that make it cheaper to run those models, despite the model itself not changing?
  2. Is the cost of performing a particular task at a particular quality level going down, through releases of newer models of similar performance (i.e., a smaller model of the current generation performing similarly to a bigger model of the previous generation, such that the cost is now cheaper)?
  3. Is the cost of running the largest flagship frontier models going down for any given task? Or does running the cutting edge show-off tasks keep increasing in cost, but where the companies argue that the improvement in performance is worth the cost increase?

I suspect that the reason why the discussion around this is so muddled online is because the answers are different depending on which of the 3 questions is meant by "is running an AI model getting cheaper over time?" And the data isn't easy to synthesize because each model has different token prices and different number of tokens per query.

But I wanted to hear from people who are knowledgeable about these topics.

 

Curious what everyone else is doing with all the files that are generated by photography as a hobby/interest/profession. What's your working setup, how do you share with others, and how are you backing things up?

view more: next ›