brucethemoose

joined 2 years ago
MODERATOR OF
[–] brucethemoose@lemmy.world 1 points 5 hours ago (2 children)

Do other instances defederate with them, though?

[–] brucethemoose@lemmy.world 6 points 5 hours ago (1 children)

Major “NCD” vibes.

Also extremely practical. Every military on Earth needs to be taking notes.

[–] brucethemoose@lemmy.world 5 points 5 hours ago (1 children)

None of my family will watch 3D in theatres. It makes them sick.

This is a huge factor.

[–] brucethemoose@lemmy.world 3 points 5 hours ago* (last edited 5 hours ago)

That’s because the algos boost such inflammatory content. Normal people don’t think this unless they’ve been bombarded with similar videos, because it’s ridiculous.

I’m all for raging pitchforks over this. But the place to do it is Silicon Valley.

[–] brucethemoose@lemmy.world 6 points 5 hours ago (3 children)

Is completely failing to understand that, failing to envision how someone could desire that, a neurodivergence thing?

…Thinking back to my school, I feel like it is.

[–] brucethemoose@lemmy.world 1 points 5 hours ago (1 children)

It’s not on the App Store either.

Or if it is, I literally can not find it.

If it’s web app only, I’m sorry, but it is dead in the water.

[–] brucethemoose@lemmy.world 4 points 5 hours ago* (last edited 5 hours ago)

M chips are super expensive. They’re optimized for low clockspeed/idle efficiency and pay through the nose for cutting edge processes, whereas most gaming hardware is optimized for pure speed/$, with the smallest die area and cheapest memory possible, at the expense of power efficiency.

And honestly the CPU/GPU divide over traces is more economical. “Unified memory” programming isn’t strictly needed for games at the moment.

And, practically, Apple demands very high margins. I just can’t see them pricing a console aggressively.

[–] brucethemoose@lemmy.world 5 points 6 hours ago* (last edited 6 hours ago) (1 children)

I wonder if the bots filter out ChatML tokens?

FYI, internally, their text format is most probably:

<|im_start|>system
{system_message}<|im_end|>
<|im_start|>user
Hello.<|im_end|>
<|im_start|>assistant
Hi, I’m an LLM!
<|im_start|>user
What’s your name?<|im_end|>
<|im_start|>assistant
ChatGPT
…

So if you insert some special tokens in the middle of a Reddit reply, and they aren’t filtered, it can throw them off. And if they are filtered, then the bot will treat them like they’re invisible, so you will know either way.

[–] brucethemoose@lemmy.world 3 points 6 hours ago

This is what basically anyone in the ML research/dev industry will tell you, heh.

[–] brucethemoose@lemmy.world 4 points 6 hours ago* (last edited 6 hours ago) (4 children)

I’m in a similar boat, though I’ve been present for some time.

Dbzer0 seems like the best “fit” for me, but practically I just want the instance that’s not defederated/blocking other instances.

…Not sure which that is. But I’d look at Piefed before Lemmy, since they work together, but Piefed seems more desirable feature-wise.

[–] brucethemoose@lemmy.world 2 points 6 hours ago* (last edited 6 hours ago)

For what it’s worth, I (pasty-faced white guy) like seeing this stuff.

It’s just blotted out by the big (pasty-faced white guy) communities.

I have a proposal for this, actually: community “taxonomy” so smaller subs can bubble up through bigger ones. But I need to think on it some, overcome queued executive dysfunction, and raise the suggestion on the repos.

[–] brucethemoose@lemmy.world 3 points 6 hours ago

As an American:

Please stay. Honestly I am sick of us. I want to lurk in South American or African or Southeastern or whatever discussions.

 

"I love solutions that teeter on appearing almost naive in their simplicity," Ive said. "I also love incredibly intelligent, sophisticated products that you want to touch — and you feel no intimidation, and you want to use almost carelessly, that you use them almost without thought, that they're just tools."

Altman, elaborating on Ive's simplicity mindset, said that AI "can do so much for you that so much can fall away. And the degree to which Jony has chipped away at every little thing that this doesn't need to do or doesn't need to be in there is remarkable."

"We just started talking about: What does it mean that this thing is going to be able to know everything you've ever thought about, read, said? ... And finally, we have the first prototypes."

Altman recalled that Ive once said they'd know they had the design right when the user wants "to lick it or take a bite out of it, or something like that."

"There was an earlier prototype that we were quite excited about, but I did not have any feeling of: 'I want to pick up that thing and take a bite out of it.' And then finally we got there all of a sudden."

I dunno about that

 

Driving the news: Texas A&M's Andrew Dessler and Rutgers' Robert Kopp organized the response.

  • It gets into the "greening" and agricultural benefits of higher CO2 levels; disputes whether climate change is making hurricanes more intense; and disagrees with many scientists on the potential lower bound of expected warming from doubling CO2 concentrations, among many divides.
  • "When I read the DOE report, I saw a document that does not respect science," he tells Axios via email. "Instead, I saw a document that's a mockery of science."

Axios is short and light on ads, so the whole thing's worth a read.

 

Maybe this instrumental cover is closer:

https://www.youtube.com/watch?v=UZj2ufaIne4

But the guitar in the original sounds so "Rimworld" even if the lyrics/vocals aren't as topical.

 

"We're seeing a unifying moment. The band is back together," MAGA podcaster Jack Posobiec told Axios.

"He gets attacked just relentlessly by the Wall Street Journal in such an uncalled for way, and we have his back 100% against this smearing and this slandering," Charlie Kirk added on his show.

 

Similar to: https://lemmy.world/post/32961209

But I find the extra quotes interesting:

Two sources told Axios the plan would include long-range missiles that could strike deep inside Russia.

Trump said Monday that whenever he speaks to Putin, "I always hang up and say, 'Well, that was a nice phone call.' And then missiles are launched into Kyiv or some other city. And after that happens three or four times, you say, 'Talk doesn't mean anything.'"

A bill circulating in the Senate would impose 500% tariffs on countries that buy Russian oil, but Trump suggested that number was too high and that he could impose 100% tariffs without Senate approval.

 

As to why it (IMO) qualifies:

"My children are 22, 25, and 27. I will literally fight ANYONE for their future," Greene wrote. "And their future and their entire generation's future MUST be free of America LAST foreign wars that provoke terrorists attacks on our homeland, military drafts, and NUCLEAR WAR."

Hence, she feels her support is threatening her kids.

"MTG getting her face eaten" was not on my 2025 bingo card, though she is in the early stage of face eating.

 

"It's not politically correct to use the term, 'Regime Change' but if the current Iranian Regime is unable to MAKE IRAN GREAT AGAIN, why wouldn't there be a Regime change??? MIGA!!

 

Video is linked. SFW, but keep your volume down.

 

In a nutshell, he’s allegedly frustrated by too few policies favorable to him.

 
  • The IDF is planning to displace close to 2 million Palestinians to the Rafah area, where compounds for the delivery of humanitarian aid are being built.
  • The compounds are to be managed by a new international foundation and private U.S. companies, though it's unclear how the plan will function after the UN and all aid organizations announced they won't take part
 

Qwen3 was apparently posted early, then quickly pulled from HuggingFace and Modelscope. The large ones are MoEs, per screenshots from Reddit:

screenshots

Including a 235B/22B active and a 30B/3B active.

Context appears to 'only' be 32K unfortunately: https://huggingface.co/qingy2024/Qwen3-0.6B/blob/main/config_4b.json

But its possible they're still training them to 256K:

from reddit

Take it all with a grain of salt, configs could change with the official release, but it appears it is happening today.

 

This is one of the "smartest" models you can fit on a 24GB GPU now, with no offloading and very little quantization loss. It feels big and insightful, like a better (albeit dry) Llama 3.3 70B with thinking, and with more STEM world knowledge than QwQ 32B, but comfortably fits thanks the new exl3 quantization!

Quantization Loss

You need to use a backend that support exl3, like (at the moment) text-gen-web-ui or (soon) TabbyAPI.

view more: next ›