BlueMonday1984

joined 2 years ago
[–] BlueMonday1984@awful.systems 7 points 2 weeks ago

trying to explain why a philosophy background is especially useful for computer scientists now, so i googled “physiognomy ai” and now i hate myself

Well, I guess there's your answer - "philosophy teaches you how to avoid falling for hucksters"

[–] BlueMonday1984@awful.systems 5 points 2 weeks ago

Its also completely accurate - AI bros are not only utterly lacking in any sort of skill, but actively refuse to develop their skills in favour of using the planet-killing plagiarism-fueled gaslighting engine that is AI and actively look down on anyone who is more skilled than them, or willing to develop their skills.

[–] BlueMonday1984@awful.systems 9 points 2 weeks ago

Someone finally flipped a switch. As of a few minutes ago, Grok is now posting far less often on Hitler, and condemning the Nazis when it does, while claiming that the screenshots people show it of what it’s been saying all afternoon are fakes.

LLMs are automatic gaslighting machines, so this makes sense

[–] BlueMonday1984@awful.systems 27 points 2 weeks ago (1 children)

Sarah Skidd, in Arizona, was called in to fix some terrible chatbot website writing. She charged $100 an hour [...] Skidd now has a side business fixing these.

The AI bros were right - AI is creating new business opportunities /s

Where there’s muck, there’s brass. And sometimes the muck is toxic waste. And radioactive. So if you get called in to fix a vibe-slopchurned disaster, charge as much as you can. Then charge more than that.

If someone's using AI, its a sign that they're (a) Nigerian Prince levels of gullible and (b) an anti-human tech asshole who fundamentally does not respect labour. Scamming these kinds of people is a moral duty.

[–] BlueMonday1984@awful.systems 17 points 2 weeks ago (2 children)

Another day, another jailbreak method - a new method called InfoFlood has just been revealed, which involves taking a regular prompt and making it thesaurus-exhaustingly verbose.

In simpler terms, it jailbreaks LLMs by speaking in Business Bro.

[–] BlueMonday1984@awful.systems 13 points 2 weeks ago (1 children)

"Another thing I expect is audiences becoming a lot less receptive towards AI in general - any notion that AI behaves like a human, let alone thinks like one, has been thoroughly undermined by the hallucination-ridden LLMs powering this bubble, and thanks to said bubble’s wide-spread harms […] any notion of AI being value-neutral as a tech/concept has been equally undermined. [As such], I expect any positive depiction of AI is gonna face some backlash, at least for a good while."

Me, two months ago

Well, it appears I've fucking called it - I've recently stumbled across some particularly bizarre discourse on Tumblr recently, reportedly over a highly unsubtle allegory for transmisogynistic violence:

You want my opinion on this small-scale debacle, I've got two thoughts about this:

First, any questions about the line between man and machine have likely been put to bed for a good while. Between AI art's uniquely AI-like sloppiness, and chatbots' uniquely AI-like hallucinations, the LLM bubble has done plenty to delineate the line between man and machine, chiefly to AI's detriment. In particular, creativity has come to be increasingly viewed as exclusively a human trait, with machines capable only of copying what came before.

Second, using robots or AI to allegorise a marginalised group is off the table until at least the next AI spring. As I've already noted, the LLM bubble's undermined any notion that AI systems can act or think like us, and double-tapped any notion of AI being a value-neutral concept. Add in the heavy backlash that's built up against AI, and you've got a cultural zeitgeist that will readily other or villainise whatever robotic characters you put on screen - a zeitgeist that will ensure your AI-based allegory will fail to land without some serious effort on your part.

[–] BlueMonday1984@awful.systems 11 points 2 weeks ago (9 children)

My only hope for this is that the GPUs in these CDO spiritual successors become dirt cheap afterwards.

They hopefully will, since the end of the AI bubble will kill AI for good and crash GPU demand.

[–] BlueMonday1984@awful.systems 12 points 3 weeks ago (1 children)

Bonus: He also appears to think LLM conversations should be exempt from evidence retention requirements due to ‘AI privilege’ (tweet).

Hot take of the day: Clankers have no rights, and that is a good thing

[–] BlueMonday1984@awful.systems 8 points 3 weeks ago (1 children)

Sidenote: The rats should count themselves extremely fucking lucky they've avoided getting skewered by South Park, because Parker and Stone would likely have a fucking field day with their beliefs

[–] BlueMonday1984@awful.systems 11 points 3 weeks ago

Apparently linkedin’s cofounder wrote a techno-optimist book on AI called Superagency: What Could Possibly Go Right with Our AI Future.

This sounds like its going to be horrible

Zack of SMBC has thoughts on it:

Ah, good, I'll just take his word for it, the thought of reading it gives me psychic da-

the authors at one point note that in 1984, Big Brother's listening device means there is two way communication, and so the people have a voice. He wonders why Orwell didn't think of this.

The closest thing I have to a coherent response is that Boondocks clip of Uncle Ruckus going "Read, nigga, read!" (from Stinkmeaner Strikes Back, if you're wondering) because how breathtakingly stupid do you have to be to miss the point that fucking hard

[–] BlueMonday1984@awful.systems 8 points 3 weeks ago

“biological civilization is about to create artificial superintelligence” is it though?

I'm gonna give my quick-and-dirty opinion on this, don't expect a lengthy defence.

Short answer, no. Long answer: no, intelligence cannot be created by blindly imitating it with mere silicon

 

Damn nice sneer from Charlie Warzel in this one, taking a direct shot at Silicon Valley and its AGI rhetoric.

Archive link, to get past the paywall.

 

(Gonna expand on a comment I whipped out yesterday - feel free to read it for more context)


At this point, its already well known AI bros are crawling up everyone's ass and scraping whatever shit they can find - robots.txt, honesty and basic decency be damned.

The good news is that services have started popping up to actively cockblock AI bros' digital smash-and-grabs - Cloudflare made waves when they began offering blocking services for their customers, but Spawning AI's recently put out a beta for an auto-blocking service of their own called Kudurru.

(Sidenote: Pretty clever of them to call it Kudurru.)

I do feel like active anti-scraping measures could go somewhat further, though - the obvious route in my eyes would be to try to actively feed complete garbage to scrapers instead - whether by sticking a bunch of garbage on webpages to mislead scrapers or by trying to prompt inject the shit out of the AIs themselves.

The main advantage I can see is subtlety - it'll be obvious to AI corps if their scrapers are given a 403 Forbidden and told to fuck off, but the chance of them noticing that their scrapers are getting fed complete bullshit isn't that high - especially considering AI bros aren't the brightest bulbs in the shed.

Arguably, AI art generators are already getting sabotaged this way to a strong extent - Glaze and Nightshade aside, ChatGPT et al's slop-nami has provided a lot of opportunities for AI-generated garbage (text, music, art, etcetera) to get scraped and poison AI datasets in the process.

How effective this will be against the "summarise this shit for me" chatbots which inspired this high-length shitpost I'm not 100% sure, but between one proven case of prompt injection and AI's dogshit security record, I expect effectiveness will be pretty high.

 

I don’t think I’ve ever experienced before this big of a sentiment gap between tech – web tech especially – and the public sentiment I hear from the people I know and the media I experience.

Most of the time I hear “AI” mentioned on Icelandic mainstream media or from people I know outside of tech, it’s being used as to describe something as a specific kind of bad. “It’s very AI-like” (“mjög gervigreindarlegt” in Icelandic) has become the talk radio short hand for uninventive, clichéd, and formulaic.

babe wake up the butlerian jihad is coming

 

I stopped writing seriously about “AI” a few months ago because I felt that it was more important to promote the critical voices of those doing substantive research in the field.

But also because anybody who hadn’t become a sceptic about LLMs and diffusion models by the end of 2023 was just flat out wilfully ignoring the facts.

The public has for a while now switched to using “AI” as a negative – using the term “artificial” much as you do with “artificial flavouring” or “that smile’s artificial”.

But it seems that the sentiment might be shifting, even among those predisposed to believe in “AI”, at least in part.

Between this, and the rise of "AI-free" as a marketing strategy, the bursting of the AI bubble seems quite close.

Another solid piece from Bjarnason.

view more: ‹ prev next ›