- Take screenshot of screen
- Send screenshot to Claude 3.7 or any other decent chatbot. Even GPT-3 is light years smarter than current Siri, but I don’t know it if can analyze images
- Run Claude’s response through a text-to-Siri-speech layer
There, now Siri can see your screen. Why can’t Apple accomplish this? Why are megacorps so inept?