Hacker News new | past | comments | ask | show | jobs | submit login

Maybe $500 billion will finally be enough to make Siri useful for more than just setting an alarm.



Funny I turned Siri off because i didn't want apple intelligence running amok. The follow-on problem --> lack of Siri killed my Carplay because Siri is required (also use itf for setting alarms/timers). The kicker? I can't seem to turn Siri back on after look through all the menus.

I.e. My preference for apple CarPlay supersedes my concerns on GPT running over my contents. Though the UI/UX has made it next to impossible to turn it back on.

What a world to live in.


Apple Intelligence and Siri are still separate (though Apple like to make it look like they are fundamentally intertwined). You can turn Apple Intelligence off and leave Siri on for CarPlay.

How did you turn Siri off in the first place? That's where I'd start...

Settings -> Apple Intelligence & Siri:

Talk and Type to Siri -> turn all this back on

Allow Siri When Locked -> turn this back on


The part of Siri that causes the most trouble is the speech recognition - which uses a voice recognition model that we now colloquially refer to as "AI." The part that works reliably, the part that sets your alarm or sends the message, is an action that's hardcoded.

IMO, moving towards AI just leads to increased uncertainty and undesirable outcomes, which is something several journalists reviewing Apple Intelligence have attested to.


If your phone is new enough for Apple Intelligence, Siri is now under that umbrella. There's no "just Siri" option anymore, unless you're rockin an iPhone 14 or older.


Or a non-Pro iPhone 15.


I would pay $500bn to get siri to distinguish between a 13 minute timer and a 30 minute timer.


I am very resourceful so I got around this by setting a 14 minute timer to outsmart the Ai.

"Your 40 minute timer starts now".


Workarounds: 781 seconds timer, or 29/31 minute timer.

I’m serious, that’s what I use.


When I put in timers -- for some reason my timer frequently/randomly just sets to 79 hours and a random assortment of minutes and seconds. I have no clue why. I always have to double check otherwise I might be waiting awhile.

It feels like it was a residual timer or something but I have never set anything like that - it is quite strange.


I solve this by asking for a 31, 41, or 51 minute timer.


How does this help me se a 13, 14, or 15 minute timer?


Could you ask to set a 780 second timer?


You can, and it works.


"10 minute timer" + "add 4 minutes"


I don't know if this is an actual problem you have, but since Siri appears to be composed of independent voice-to-text and text-to-action systems, you can say "start a one three minute timer".


The problem is AI current best use case is creative work, art, music, programming, but skilled creative professionals is a/the core userbase for Apple products.

Apple is stuck and it’s AI will never be good enough until those creatives embrace it. Right now it’s disdain when mentioned.


> The problem is AI current best use case is creative work, art, music, programming

This is where it’s being pushed and marketed but I’m not actually sure it’s the best use case.


An oft-cited quote goes something like this: "we wanted robots/AI to automate boring, routine, meaningless jobs to let people be free to pursue arts, music, creativity. It's a sad state of affairs that AI is taking over arts/music/creativity stranding people with boring, routine, meaningless jobs"


> It's a sad state of affairs that AI is taking over arts/music/creativity stranding people with boring, routine, meaningless jobs"

So far it’s not though.


Oh. It already is. Artists are already saying that a lot of commission work is drying out (e.g. illustrations).


> skilled creative professionals is a/the core userbase for Apple products.

Then why doesn't it have a professional CAD application?


Even Apple doesn't do CAD on OSX. They run Siemens NX on Windows.

But your statement isn't quite right. Fusion 360 runs fine in mac. I'm ex-Apple btw.


Ok, that's good to know, but my colleagues all use SolidWorks so it doesn't change much for me.


It does.

AutoCAD came to the Mac when Intel was shitting the bed (with aggressive OEM contracts for first party system integrators that prevented AMD adoption across HP/Dell/Lenovo-lines) and Windows 11 was being forced on users.

WINTEL played the monopoly game too hard and is starting to lose ground.

You love to see it.


Uhh

> The problem is AI current best use case is creative work, art, music, programming

By “best” do you mean “marketable?”

Seems weird to see a bunch of creatives and creative professionals “disdain” a tool and still say it’s “best” for them…


Even that seems to work only half the time. ~50% she just doesn't respond to a voice command "Siri" or "Hey Siri" for me.


When Siri first debuted it would automatically beep, so I could immediately tell if the phone did not recognize recognize "Hey Siri" (just "Siri" didn't work). A couple of iOS updates later this went away, which means I can't tell without actually picking up the phone and looking at it whether the command was accepted.

Even more annoyingly, sometimes there is a beep! ¯\_(ツ)_/¯


Yup, the UX has gotten so much worse.


Whoa, whoa, whoa, on occasion, if the planets align just right, I can also get Siri to set a reminder (and at least half the time Siri gets it 80% right).

LLM Siri cannot come fast enough.


If you haven't tried OpenAI's advanced voice mode, it's a mind blowing version of exactly what things like Siri really ought to become with a little more development. If that's what you mean by LLM Siri, I totally agree.

Being able to chat casually with low latency, correct yourself, switch languages mid-sentence, incorporate context throughout a back-and-forth conversation etc. turns talking to these kinds of systems from a painful chore into something that can actually add value.


But does it do actual stuff, like adding a meeting to your calendar, call people, or set a timer or a reminder?


LLMs are there for generating/hallucinating text, not for understanding text.

For natural language processing you need a different kind of neural network don't you?


It's the other way around. The model is impeccable at "understanding text." It's a gigantic mathematical spreadsheet that quantifies meaning. The model probably "understands" better than any human ever could. Running that backwards into producing new text is where it gets hand-wavy & it becomes unclear if the generative algorithms are really progressing on the same track that humans are on, or just some parallel track that diverges or even terminates early.


I thought it quantifies the probability that a certain word (their output) follows a given word sequence (their training corpus and the prompt)?


Only if you wildly oversimply to the level of being misleading.

The precise mechanism LLMs use for reaching their probability distributions is why they are able to pass most undergraduate level exams, whereas the Markov chain projects I made 15-20 years ago were not.

Even as an intermediary, word2vec had to build a space in which the concept of "gender" exists such that "man" -> "woman" ~= "king" -> "queen".


> Only if you wildly oversimply to the level of being misleading.

Maybe I'm asking for an explanation :)

Since you seem to understand the mechanism, can you do a 3 line summary please?


3 lines? That's still going to be oversimplifed to the point of being wrong, but OK.

Make a bunch of neural nets to recognise every concept, the same way you would make them to recognise numbers or letters in handwiting recognition. Glue them together with more neural nets. Put another on the end to turn concepts back into words.

For a less wrong but still introductory summary that still glosses over stuff, about 1.5 hours of 3blue1brown videos, #4-#8 in this playlist: https://youtube.com/playlist?list=PLZHQObOWTQDNU6R1_67000Dx_...


Thanks!

... Oh interesting. And those concepts are hand picked or generated automatically somehow?

> For a less wrong but still introductory summary that still glosses over stuff, about 1.5 hours of 3blue1brown videos

Sorry, my religion forbids me from watching talking heads. I'll have to live with your summary for now. Until I run into someone who condensed those 1.5 hours in text that takes at most 30 min to read...


> Oh interesting. And those concepts are hand picked or generated automatically somehow?

Fully automated.

> Sorry, my religion forbids me from watching talking heads.

What about professional maths communicators who created their own open sourced python library for creating video content and doesn't even show their face on most videos?


My problem is with the time wasted compared to written info, not with talking heads per se.


He doesn't waste time. No fluff.

You're unlikely to get a better time-quality trade-off on any maths topic than a 3blue1brown video.

He's the kind of presenter that others try to mimic because he's so good at what he does — you may recognise the visuals from elsewhere because of the library he created[0] in order to visualise the topics he was discussing.

[0] https://docs.manim.community/en/stable/faq/installation.html

There's also a playback speed slider in YouTube. I use it a lot.


Simplifying to that point is more of what a Markov chain is. LLMs are able to generalize a lot more than that, and it's sufficient to "understand text" on a decent level. Even a relatively small model can take, e.g. even this poorly prompted request:

  "The user has requested 'remind me to pay my bills 8 PM tomorrow'. The current date is 2025-02-24. Your available commands are 'set_reminder' (time, description), 'set_alarm' (time), 'send_email' (to, subject, content). Respond with the command and its inputs."
And the most likely response will be what the user wanted.

A Markov chain (only using the probabilities of word orders from sentences in its training set) could never output a command that wasn't stitched together from existing ones (i.e. it would always output a valid command name, but if no one had requested a reminder for a date in 2026 before it was trained, it would never output that year). No amount of documents saying "2026 is the year after 2025" would make a Markov chain understand that fact, but LLMs are able to "understand" that.


I’m confident that LLM’s will not have hallucination problems in the type of requests that I send to Siri.

I don’t ask Siri for facts (just like I don’t ask LLM’s for facts). As long as it can correctly, understand what and when I ask to be reminded about something, that would be a huge improvement for me.

That and being able to map “Bedroom Fan”/“Bedroom Fan Light” to “Bedroom Fan Lights” without having to specify aliases (and even then it hearing me wrong).

I’ve see Home Assistant working with LLMs and it can understand groupings that I never explicitly defined which is very nice. I can say “Turn off all overhead lights” and it will find all my overhead lights and turn them off. Siri/Alexa can’t handle those tasks currently.


They are AMAZING at understanding. I think even more reliably than generating. And with context, and back and forth.

Try asking ChatGPT to remember some obscure film you can only remember very hazy details about - really random stuff - I bet it will identify it for you I a few tries.

You can also totally miss spell words, use messed up grammar and it has no problems at all


What are you talking about? We've invented the Universal Translator already.


> LLM Siri cannot come fast enough.

I don't think it's been demonstrated that Apple could make Siri better with an LLM.


Over on Android it's the opposite situation. The voice interface to Google Assistant was very reliable for simple things like reminders and appointments, and even for general knowledge questions. It was part of why I didn't switch to an iPhone. Then Gemini came along, and that core functionality got a lot worse.


Siri already has LLM integration...

...that will grind your request to set email Vacation Mode through the world's worst speech-to-text, jam the text into Chat GPT, and spend the next three minutes reading you an uninterruptible 3 minute essay about violence.


I tried this with the new ‘apple intelligence’ that I thought could see my screen.

I had a birthday invite with clear date and time: so I asked it to add to my calendar.

It just said, “add what?” repeatedly until finally deciding it needed to send it to chatGPT to help. Which it did, then just returned the text in the image without taking any action. Then I say “can you add the event now?”

“Add what?”

So I try copying the text from the image in photos and giving that to Siri to add as an event. Surly this can work?!?

Nope.

It’s just utterly pitiful.


I think that feature has been delayed.


I'm a formerly non-mac guy who finally bought a brand new iPad. I got bored with the wallpaper but couldn't figure out how to change it. "Hey Siri, how do I change the wallpaper?"..."Sorry, I can't help with that". Tried a couple more questions and all it did was Google it for me. This is the latest M4 that was around $2k.

This is what our "AI accelerated" chips give us in return? What a disgrace


Apple software is already largely written my Americans.

Somehow I don't think fealty will change its quality.


hey thats what alexa is for


No, Alexa is only good to interrupt you randomly when you're watching TV.




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: