this post was submitted on 28 Aug 2025
22 points (89.3% liked)

Technology

40118 readers
259 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 3 years ago
MODERATORS
 

As always, I use the term "AI" loosely. I'm referring to these scary LLMs coming for our jobs.

It's important to state that I find LLMs to be helpful in very specific use cases, but overall, this is clearly a bubble, and the promises of advance have not appeared despite hundreds of billions of VC thrown at the industry.

So as not to go full-on polemic, we'll skip the knock-on effects in terms of power-grid and water stresses.

No, what I want to talk about is the idea of software in its current form needing to be as competent as the user.

Simply put: How many of your coworkers have been right 100% of the time over the course of your career? If N>0, say "Hi" to Jesus for me.

I started working in high school, as most of us do, and a 60% success rate was considered fine. At the professional level, I've seen even lower with tenure, given how much things turn to internal politics past a certain level.

So what these companies are offering is not parity with senior staff (Ph.D.-level, my ass), but rather the new blood who hasn't had that one fuckup that doesn't leave their mind for weeks.

That crucible is important.

These tools are meant to replace inexperience with incompetence, and the beancounters at some clients are likely satisfied those words look similar enough to pass muster.

We are, after all, at this point, the "good enough" country. LLM marketing is on brand.

you are viewing a single comment's thread
view the rest of the comments
[–] BlameThePeacock@lemmy.ca 2 points 1 day ago (6 children)

I just implemented an LLM in a vacation request process precisely because the employees are stupid as fuck.

We were getting like 10% of requests coming in with the wrong number of hours requested because people can't fucking count properly, or understand that you do not need to use vacation hours for statutory holidays. This is despite the form having a calculator and also showing in bright red any stat holidays inside the dates of the request.

Now the LLM checks if the dates, hours, and note from the employee add up to something reasonable. If not it goes to a human to review. We just had a human reviewing every single request before this, because it was causing so many issues, an hour or two each week.

[–] lucas@startrek.website 9 points 1 day ago* (last edited 1 day ago) (1 children)

Why would you use an LLM for this? This sounds like a process easily handled by conventional logic, which would be cheaper, faster, and actually reliable... (The 'notes' part notwithstanding I guess, but calculations in general are definitely not a good use of an LLM)

[–] BlameThePeacock@lemmy.ca 2 points 1 day ago (1 children)

Normally I'd agree, and we used some of that in the original form (like maximum hours, checking for negative submissions, etc.) but requests don't always follow simple logic and more complex logic just led to failures every time a user did something other than take a standard full day off.

Some employees work 7 hours, while others are 7.5, some have flex days and hours that change that, sometimes requests are only for part days, sometimes they may use multiple leave types to cover one off period.

I spent a few hours writing and testing the prompt against previous submissions to fine tune it.

So far it's detected every submission error in the two weeks it's been running, with only one false positive.

[–] jbloggs777@discuss.tchncs.de 3 points 1 day ago

If it helps to accurately fill in the details correctly in the backend system, which are then properly validated or escalated for human review/intervention (and let the human requester choose the escalation path too, as opposed to blindly submitting), then it sounds great.

Guided experiences, leading to the desired outcome, with less need for confused humans to talk to confused humans.

I want the same for most financial approvals in my company. Finance folks speak a different language to most employees, but they have an oversized impact on defining business processes, slowing down innovation, frustrating employees, and often driving costs UP.

load more comments (4 replies)