It's not always perfect, but it's good for getting a tldr to see if maybe something is worth reading further. As for translations, it's something AI is rather decent at. And if I go from understanding 0% to 95%, really only missing some cultural context about why a certain phrase might mean something different from face value, that's a win.
You can do a lot with AI where the cost of it not being exactly right is essentially zero. Plus, it's not like humans have a great track record for accuracy, come to think of it. It comes down to being skeptical about it like you would any other source.
At least, the iPhone notifications summaries were bad enough I eventually turned them off (but periodically check them) and while I was working at Google you couldn't really turn of the genAI summaries of internal things (that evangelists kept adding to things) and I rarely found them useful. Well... they're useful if the conversation is really bland but then the conversation should usually be in some thread elsewhere, if there was something important I don't think the genAI systems were very good at highlighting it
iPhone notification summaries were made with GPT3.5 I believe (maybe even the -turbo version).
It doesn't use reasoning and so when using very short outputs it can produce wild variations since there are not a lot of previous tokens in order to direct the LLM into the appropriate direction in kv-space and so you're more at the whims of temperature setting (randomly selecting the next token from a SOFTMAX'd list which was output from the LLM).
You can take those same messages and plug them into a good model and get much higher quality results. But good models are expensive and Apple is, for some reason, going for the budget option.
Not OP, but...
It's not always perfect, but it's good for getting a tldr to see if maybe something is worth reading further. As for translations, it's something AI is rather decent at. And if I go from understanding 0% to 95%, really only missing some cultural context about why a certain phrase might mean something different from face value, that's a win.
You can do a lot with AI where the cost of it not being exactly right is essentially zero. Plus, it's not like humans have a great track record for accuracy, come to think of it. It comes down to being skeptical about it like you would any other source.
Strongly disagree with the TLDR thing
At least, the iPhone notifications summaries were bad enough I eventually turned them off (but periodically check them) and while I was working at Google you couldn't really turn of the genAI summaries of internal things (that evangelists kept adding to things) and I rarely found them useful. Well... they're useful if the conversation is really bland but then the conversation should usually be in some thread elsewhere, if there was something important I don't think the genAI systems were very good at highlighting it
iPhone notification summaries were made with GPT3.5 I believe (maybe even the -turbo version).
It doesn't use reasoning and so when using very short outputs it can produce wild variations since there are not a lot of previous tokens in order to direct the LLM into the appropriate direction in kv-space and so you're more at the whims of temperature setting (randomly selecting the next token from a SOFTMAX'd list which was output from the LLM).
You can take those same messages and plug them into a good model and get much higher quality results. But good models are expensive and Apple is, for some reason, going for the budget option.
AFAIK some outputs are made with a really tiny/quantized local LLM too.
And yeah, even that aside, GPT 3.5 is really bad these days. It’s obsolete.