← Back to context

Comment by yalogin

3 days ago

They nailed it. Consumers don't care about AI, they care about functionality they can use, and care less if it uses AI or not. It's on the OS and apps to figure out the AI part. This is why even though people think Apple is far behind in AI, they are doing it at their own pace. The immediate hardware sales for them did not get impacted by lack of flashy AI announcements. They will slowly get there but they have time. The current froth is all about AI infrastructure not consumer devices.

The only thing Apple is behind on in the AI race is LLMs.

They've been vastly ahead of everyone else with things like text OCR, image element recognition / extraction, microphone noise suppression, etc.

iPhones have had these features 2-5 years before Android did.

  • Apple’s AI powered image editor (like removing something from the background) is near unusable. Samsung’s is near magic, Google’s seems great. So there’s a big gap here.

    • That is rather funny because I think Google's and Samsung's AI image actions are completely garbage, butchering things to the point where I'd rather do it manually on my desktop or use prompt editing (which to Google's credit Gemini is fantastic at). Whereas Apple's is flawless in discerning everything within a scene or allowing me to extract single items from within a picture. For example say, a backpack in the background.

    • > unusable

      apple is so hit or miss.

      I think the image ocr is great and usable. I can take a picture of a phone number and dial it.

      but trying to edit a text field is such a nightmare.

      (try to change "this if good" to "this is good" on iphone with your fingers is non-apple cumbersome)

    • Well if I ever used an slop-image-generator, that’d be an issue, but as I don’t, it’s a bit of a non-event!

  • > had these features 2-5 years before Android did.

    "first" isn't always more important than "best". Apple has historically been ok with not being first, as long as it was either best or very obviously "much better". It always, well, USED TO focus on best. It has lost its way in that lately.

  • TTS is absolutely horrible on iOS. I have nearly driven into a wall when trying to use it whilst driving and it goofs up what I've said terribly. For the love of all things holy, will someone at Apple finally fix text to speech? It feels like they last touched it in 2016. My phone can run offline LLMs and generate images but it can't understand my words.

    • > I have nearly driven into a wall when trying to use it whilst driving and it goofs up what I've said terribly.

      People should not be using their phones while driving anyways. My iPhone disables all notifications, except for Find My notifications, while driving. Bluetooth speaker calls are an exception.

  • Kind of a big "only" though. Siri is still shit and it's been 15 years since initial release.

    • When I'm driving and tell Siri, "Call <family member name>", sometimes instead of calling, it says, "To who?", and I can't get it to call no matter what I do.

    • Amazing how its been 15 years and it still can't discern 15 from 50 when you talk to it.

> did not get impacted by lack of flashy AI announcements

To be fair, they did announce flashy AI features. They just didn't deliver them after people bought the products.

I've been reading about possible class action lawsuits and even the government intervening for false advertisement.

All of the reporting about Apple being behind on AI is driving me insane and I hope that what Dell is doing is finally going to be the reversal of this pattern.

The only thing that Apple is really behind on is shoving the word (word?) "AI" in your face at every moment when ML has been silently running in many parts of their platforms well before ChatGPT.

Sure we can argue about Siri all day long and some of that is warranted but even the more advanced voice assistants are still largely used for the basics.

I am just hoping that this bubble pops or the marketing turns around before Apple feels "forced" to do a copilot or recall like disaster.

LLM tech isn't going away and it shouldn't, it has its valid use cases. But we will be much better when it finally goes back into the background like ML always was.

  • Right! Also I don’t think Siri is that important to the overall user experience on the ecosystem. Sure it’s one of the most visible use cases but how many people really care about that? I don’t want to talk out loud to do tasks usually, it’s helpful in some specific scenarios but not the primary use case. The text counterpart of understanding user context on the phone is more important even in the context of llms, and that what plays into the success of their stack going forward

    • are you really asking why someone would like a much better siri?

      - truck drivers that are driving for hours.

      - commuters driving to work

      - ANYONE with a homepod at home that likes to do things hands free (cooking, dishes, etc).

      - ANYONE with airpods in their ears that is not in an awkward social setting (bicycle, walking alone on the sidewalk, on a trail, etc)

      every one of these interaction modes benefits from a smart siri.

      That’s just the tip of the iceberg. Why can’t I have a siri that can intelligently do multi step actions for me? “siri please add milk and eggs to my Target order. Also let my wife know that i’ll pick up the order on my way home from work. Lastly, we’re hosting some friends for dinner this weekend. I’m thinking Italian. Can you suggest 5 recipes i might like? [siri sends me the recipes ASYNC after a web search]”

      All of this is TECHNICALLY possible. There’s no reason apple couldn’t build out, or work with, various retailers to create useful MCP-like integrations into siri. Just omit dangerous or destructive actions and require the user to manually confirm or perform those actions. Having an LLM add/remove items in my cart is not dangerous. Importantly, siri should be able to do some tasks for me in the background. Like on my mac…i’m able to launch Cursor and have it work in agent mode to implement some small feature in my project, while i do something else on my computer. Why must i stare at my phone while siri “thinks” and replies with something stupid lol. Similarly, why can’t my phone draft a reply to an email ASYNC and let me review it later at my leisure? Everything about siri is so synchronous. It sucks.

      It’s just soooo sooo bad when you consider how good it could be. I think we’re just conditioned to expect it to suck. It doesn’t need to.

      2 replies →

Even customers who care about AI (or perhaps should...) have other concerns. With the RAM shortage coming up many customers may choose to do without AI features to save money even though they want it at a lower price.

Nailed it? Maybe close. They still have a keyboard button dedicated to Copoilot. That thing can’t be reconfigured easily.

  • Can PowerToys remap it?

    • Yes, on my Thinkpad I could remap it with Powertoys. It looks like the sibling comments have had issues though.

      For me, the Copilot key outputs the chord "Win (Left) + Shift (Left) + F23". I remapped it to "Ctrl (Right)" and it's functioning as it should.

    • I have one laptop with a Copilot key in my business. (I didn't even realize that when I bought it.) It takes the place of a modifier key, I think the menu key. Except it outputs a specific keypress (Ctrl+Shift+F23). So it can't be mapped to anything useful like a modifier key. But you can reassign the meaning of Ctrl+Shift+F23.

    • You can sort of remap it on windows, but it's somewhat limited in my experience. It shows up as a keyboard chord rather than a simple button press. I think it's LWin+LShift+F23. I ended up simply disabling it entirely on my gaming laptop. I've been meaning to see if it's easier to make it useful on KDE Plasma desktop but haven't yet (though I did remap the HP Omen button to pull down Yakuake instead).