Personal theory of mine is
*.itjust.works
meant to stand for "It Just Works" until they decided to give this Lemmy thing a go.
Yep it's referencing a meme that originated almost a decade ago. https://youtu.be/nVqcxarP9J4
Personal theory of mine is
*.itjust.works
meant to stand for "It Just Works" until they decided to give this Lemmy thing a go.
Yep it's referencing a meme that originated almost a decade ago. https://youtu.be/nVqcxarP9J4
Otan varmaan jossain välissä, nyt kun muuttanut pois asuinalueelta jossa uusi sähköpyörä olisi varmasti lähtenyt kuukauden sisään kävelemään. Toivoisin kyllä edun uudelleen miettimistä niin, että olisi yhtä lailla kannattava kaikille. Onhan tuo nyt vähän perverssiä, että minulle hyvitetään pyörästä lähemmäs puolet mutta 2000/kk tienaavalle vain neljännes.
Not quite ELI5 but I'll try "basic understanding of calculus" level.
In very broad terms, the model learns complex relationships between words (or tokens to be specific, explained below) as probabilistic scores. At its simplest, this could mean the likelihood of one word appearing next to another in the massive amounts of text the model was trained with: the words "apple" and "pie" are often found together, so they might have a high-ish score of 0.7, while the words "apple" and "chair" might have a lower score of just 0.2. Recent GPT models consist of several billions of these scores, known as the weights. Once their values have been estabilished by feeding lots of text through the model's training process, they are all that's needed to generate more text.
Without getting into the math too much, this is how a GPT model then uses these numbers to come up with words:
In reality we're not quite so sure what the weights represent to the model exactly, but this is the gist of it. All we know is that they signify the importances or non-importances that the model places on some pattern that was present in the training data. Some of these patterns could be just simple two-word pairs, but many are probably much more complicated. Lots of researchers are currently trying to get a better idea of how these numbers are actually affecting the model's output.
I'm currently maintaining a multi million line VB.NET code base, the foundations of which were hastily laid down by young and inexperienced devs realizing a business opportunity in the early 2000s. Lots of these out there in the enterprise world from what I hear and I think this is where there the language gets its reputation from. Sure, at its best it's just C# with words in place of curly braces, but that's only the case with well disciplined programmers (and even then, why not just use C#?). Option Strict is, well, just an option, and even the infamous On Error Resume Next
is still usable in VB.NET to this day afaik. A lot more room for shooting yourself (or the next person reading your code) in the foot if you don't know what to look out for.
Furthermore, even if you wanted to operate an instance on a small scale you'd still have to deal with the full volume of posts from the rest of the Lemmyverse getting pulled and saved to your instance. If we had Reddit levels of activity here, every instance host with more than a couple dozen users would basically end up maintaining their own personal database copy of Reddit (more or less, provided those users were still joining the popular communities across the 'verse) which doesn't sound like something I'd want to deal with as a hobbyist.
I agree with your sentiment, but I wouldn't call activism (and especially not journalism) a wasted effort in that regard. Bringing issues to light is the first step in creating a visible dent in the balance sheets. Public perception shapes consumer behavior to some degree and can put pressure on lawmakers to introduce legislation against harmful conduct. On the other hand, if the general public only hears the company's side of the story underlining how clean and ethical they are, there will never be any pressure for change.