Gruber's blog post criticizes Apple's Siri for inconsistent results when asked 'What time is it in London?', comparing it unfavorably to Google, DuckDuckGo, Alexa, and Bing. The post exemplifies freedom of expression (Article 19) through public, unrestricted criticism of a major corporation and features an accessible publication platform (RSS feed, no paywall). No other human rights provisions are engaged.
Same thing happens with Frankfurt. If I order something from Amazon.de, my package usually goes through Frankfurt. When I checked my package tracker app on my iPhone, I was surprised to find out that Frankfurt is actually on the Germany-Poland border.
Turns out that if I enter "Frankfurt, Germany" into Apple Maps (which I assume is what the package tracker app does), it takes me to "Frankfurt (Oder), Germany" instead of "Frankfurt am Main, Germany".
Ranking and NLP aren't easy. If you are asking a slightly related question (for example, "What is the weather in London"), and if you are living at some place where the nearest major town is called London, but is not London in England, you would expect it to give you the weather in "your" London. However, it you are asking for the time in a particular city, then the ranking should of course consider whether the timezone of the city you asked for is different than your own - it makes no sense to ask for the time in a city which lies in your timezone. Then again, if the distance from your location to that other London is greater than a certain threshold, the question could imply that you actually do not know whether the city lies in the same timezone as your location.
All these thresholds or ranking factors seem to come intuitively to humans (I would guess a good intuition for them is actually a sign of intelligence), but it seems to be incredibly hard to capture them in ranking.
As others have pointed out, a solution here would be to make Siri more conversational. A simple "Which London?" could've removed the ambiguity and given Siri the opportunity to learn something about that particular person (that London, England is more important to him than London in Canada).
To throw a tongue-in-cheek additional example in here: the population of London, Ontario is (significantly) greater than the population of the City of London, UK.
Reality is hard. And with machine learning (especially proprietary, remotely-hosted machine learning) there's rarely a way to pinpoint a line of code and say: "this is what happened and why you're now frustrated and firing hypothetical personal assistants".
It's ridiculous how Siri is still this shitty. I have an 11 Pro and even on such an expensive phone I can't really trust it to do anything more advanced than set timers. Every few months I try to do something else and just get annoyed at how bad it is.
Before lockdown I even had it disabled entirely because it would get activated randomly from time to time, even if nobody in the vicinity said anything remotely close to "Hey Siri".
An actual assistant would have a lot of context. Do I know that you're going to travel to London, UK for a break next week and therefore I would naturally assume that you are interesting in the time there.
However, are you planning to visit your parents in London, Canada this weekend? Then an assistant who would still answer with the time of London, UK would maybe also not be the smartest?
So really context is everything and making broad statements that if an assistant was to answer with anything but London, UK should get fired is something that someone would say, who IMHO should get fired. shrug
Also, IMHO, if someone doesn't know that machines don't have human context and therefore doesn't know to ask their digital assistant "What is the time in London, UK" when they want to know the time in London, UK, then maybe they should get fired from their tech job. shrug
This example illustrates how difficult AGI is and how far we are from it. We, humans, tend to take advantage of the context to make communication simpler and shorter. Just think about all the implications of this one simple question: what time is it in London? Or e.g. how can I get from London to Dublin?
If the person asking the question lives in Ohio, they may actually be talking about London OH (or Dublin OH). Some people in neighbouring states may mean the same, though they will be more likely to mention the state. However, how close should you be to London, OH even within the state to mean the Ohio one and not the UK one? How close is close enough? Is a few hours of driving close enough? A 3 hour flight? What if I'm roughly at 6 hours from London OH and 7 hours from London UK?
Further, if the person is a British expat in Ohio, especially if they are working for a multinational business (or not), they would more likely mean London UK. German expats, though? Russians? Or an Irish person who lives in Amsterdam having some relatives in Ohio US, looking to book a flight to Dublin. Etc. etc.
There are so many contextual layers here that even human assistants can occasionally get it wrong, and without the context the task becomes insurmountable for the "AI" algorithms. That is not to say virtual assistants are useless, just that selling them as "AI" is a big lie, bigger than even those who market these algorithms as "AI" think it is.
For a related example, type "11:00 EST to UTC" into Google and DuckDuckGo. Google says 15:00 because it interprets "EST" colloquially as "the current time on east coast US". DuckDuckGo says 16:00 because it interprets EST literally as "the time on east coast when it isn't daylight savings time" (compared to ET or EDT). It isn't clear which behavior is more desirable.
I now have most of my adult years learning how to construct a phrase that gets the right results for Google, and later Assistant. It got to the point where I'm certain it must be a headache for whatever team is trying to support natural language processing in these - all proficient users ask for some artificial gibberish and get where they want to be.
Here comes my favourite brain freeze moment - recently my parents asked me to explain this to them. How do you construct a good search phrase? My brain blanked. I HAVE NO IDEA. It seems I have learned fluent Goonglish without noticing, and now can't explain the grammar or vocabulary of it.
Aza Raskin's Ubiquity was such a clear model of how to build voice interfaces the right way, and it wasn't even a voice interface. It was a bit of a launcher that tied APIs together on the web.
Let users create and share small commands. Create a simple natural language for commands that are easy to program, extend, and remember, and narrow the scope of inputs the voice engine has to deal with.
It was so beautiful and effective and just light years ahead of what we're getting.
Microsoft gets a special mention for lost potential here. Their voice system in Windows could be a way to navigate the layered menus of the OS, but it is mostly focused on answering general queries. Voice is a great replacement for the program launcher, except it's not customizable, but that's about the extent of how much you can control the system with it. Let me do anything buried in the control panel, show me everything you know about a process when I ask, solve that first, then worry later about telling me how big the moon is. You make an OS, don't forget what that is.
TBH I always feel amazed about how worked up people get about stuff like this, especially people familiar with software who should know that there are millions, maybe billions of edge cases like this in a generic knowledge system, and thus at least it's easy to make a mistake like this. I mean, the time it took him to write his blog post is probably more than all the times it would take him to follow up with "What time is it in London, England?" It reminds me of someone who commented that "there must not be any black people who work at Apple" because it pronounced "Malcolm X Blvd" as Malcolm 10 Blvd.
I mean, if anything, just appreciate how amazing humans are at differentiating these ambiguities.
Interesting. I can see how the basic algorithm might go "I've been asked for the time (or temp or whatever) in cityname. Citynames are routinely reused globally; what is the closest such cityname?"
But that fails completely when you get to names like London (or Paris or Moscow or Cairo).
But it happens with people, too. I'm from Mississippi, though I haven't lived there since I left for college. I now live in Houston. At a family reunion many years ago, I ran into a cousin I hadn't seen since we were kids. She asked where I was living, and I told her.
"Oh, isn't it terrible about that wreck?" she asked.
Baffled, I asked for more information. "Oh, you know, that wreck over on 406!"
I did not know. "I'm sorry, Houston's really huge. I don't know what wreck you mean."
"Oh, did you mean you live in Houston, TEXAS? I thought you meant Houston, MISSISSIPPI!"
I was, at the time, about 30. I grew up in that state, and lived there until I went to college. And until that moment, I had never even HEARD of Houston, Mississippi (a metropolis, it turns out, of about 3600 people in the misbegotten northeast corner of the state).
The craziest and most confusing behavior of Siri for me is:
Sometimes you can ask a question and watch it be perfectly transcribed in real time, but then receive a nonsensical answer from it. Ask the exact same question immediately after on the same device, transcribed exactly the same way, and get the correct answer.
Where does such unpredictability come from? How can Siri transcribe the words correctly but fail to deliver the right answer?
At some point about a year ago I noticed that I could no longer ask my Google Home devices "what's the weather?". I'd just get a generic "I don't understand" response. But more specific queries such as "What's the weather in Seattle?" would work.
After a couple of weeks of this, I somehow got the idea that it was related to the devices' configured locations. And sure enough, telling the Home that I lived in the next city over fixed the problem.
So I started a binary search and eventually found that the issue was limited to my ~10x20 block Seattle neighborhood - basically the outline shown when I search for its name in Google Maps. I then also realized that it applied to weather queries on my phone as well, but since the phone uses GPS rather than a specific location setting, I could only reproduce the broken and working behaviors by crossing one of the neighborhood boundary streets.
Turns out it was some long-standing configuration issue with Knowledge Graph's entry for my neighborhood, and some recent code change in location-based weather queries began butting heads with it. Luckily I worked at Google at the time and was able to track down and pester people that could help fix the issue.
90% of the time works fine, and it's essentially all I use Siri for. It's very convenient when cooking and my hands are dirty. But 10% of the time I get something along the lines of...
"I'm sorry, but you don't have the Timer app installed".
"I'm sorry, but you don't have the Timer app installed".
"I'm sorry, but you don't have the Timer app installed".
"I'm sorry, but you don't have the Timer app installed".
It's infuriating because I know Siri is dumb so I use the same exact simple phrases to avoid confusion. Sometimes it works, sometimes it doesn't. It always transcribes the command accurately though! I've actually lost my temper and smashed an Apple Watch before over this. This is in my house, on a very reliable network, always with my phone within a reasonable distance.
Siri doesn't know that my front door is called "FRONT DOOR".
I only have one smart lock, which works perfectly, and it is called "FRONT DOOR" in HomeKit.
When I ask Siri about my FRONT DOOR she responds that she cannot find it.
When I ask Siri about the status of my DOOR, she responds with "The FRONT DOOR is locked/unlocked".
I'll then say 'Alright Siri you literally just used the phrase "FRONT DOOR" five seconds ago and the text transcript on the screen says "FRONT DOOR" hey Siri is my FRONT DOOR locked'
Siri: WTF are you talking about? You don't have a FRONT DOOR.
There's a Woodland Hills in Utah and a Woodland Hills in California. If you ask Google what the weather is in Woodland Hills, it will ALWAYS give you the weather for California. Even if your current location is Woodland Hills, Utah and even if your address is set to Woodland Hills Utah in your Google account or Google home.
Once I had a new assistant and I asked her to book me a flight to Boston. She went to the travel booking system, typed in Boston, then called me back confused. "Which Boston do you want? There are 8 of them?"
I was caught off guard and told her I'd prefer the one in Massachusetts. I did not fire her. She was young, had a poor general education, and had never traveled outside her home state. Those things do not make her stupid.
Might as well comment, first employee at Siri here. This result “maybe” should provide ambiguity resolution, but where does it stop. The she/he who compared it to Google was right on. Siri provides singular results in most cases vs multiple search style results. We did use geo for locality based results in the past. This would solve the problem the OP mentioned, not sure if they call location for these requests now. The other person who mentioned we can’t/couldn’t train on data is correct too. Again, privacy first. Be proud, Apple cares a lot. When the Siri commercials hit (No one told us there would be commercials) when we launched, we got decimated, and couldn’t debug the issues, user utterances were not allowed to be logged. Luckily, after much sleep deprivation, one of my engineers (love you Stu) said, “hey, aren’t they running commercials” to all our surprise. We convinced the privacy team to let us log word parts. Then we started to see words that were present in the commercials. Fun fact, also happened when Tom Cruise was presenting at the Academy Awards. We had millions+ asks all at the same time, again word parts. “height”, “tom”, “foot”, etc.
My company (based in Frankfurt am Main, Germany) had an online business travel reservation system which helpfully pre-filled the starting point of all new trips to Frankfurt, Kentucky (which is not even the right spelling of the city).
I filed a request to NOT pre-fill the starting city to a place in Kentucky and got a polite but firm reply that this was the default list provided by the 3rd party online booking engine and cannot be customized. Sigh.
I’ve had Siri disabled for years. Even the basic “call home” works every 3rd time. I try it for a few minutes with every new iOS update only to see it’s still the same dumbster fire.
The problem sounds a little like collaborative filtering. If you have a certain affinity with cities A,B,C, then you can compute the expected affinity with a city X by looking at the affinities other people have with X, and their affinities with A,B,C.
Instead of looking at people, you can also scrape websites to get the relations. But here you may get a recursive problem because if a website speaks of "London", you might not know in advance which London they speak of.
Yes, but in the same way people don't usually mean "London, Ontario" when they say London, they also don't usually mean the City of London (which, for the benefit of people who may not know, is a tiny portion of London with a population less than 10,000).
ha! I was wondering how hard it'd be to find someone making an appeal to "context". While you're not exactly wrong, you are. Parsing meaning is something humans are surprisingly good at, and trained for.
I think the most damning part is how, at the bottom, he list a handful of other "smart" assistants which correctly list London UK's time... for now.
But, his point about consistency and slowness is exactly why I never use these shitty voice assistants. If I'm going to be interacting with some pedantic robot, I generally want to be able to edit the text of my request.
If you ever cross the border from Frankfurt/Oder, the first thing you see in Poland is a giant Amazon building. Despite that building being there, Polish people still can't order at Amazon.
I had never used an Apple product before the company which I joined recently gave me a MacBook Pro. I am really surprised how bad the product quality is. The calendar notification is very random. Sometimes it fires, sometimes it does not. I have missed couple of meetings because notification popped after the meeting was over. Similarly the keyboard shortcut is random. Sometimes it opens the app, sometimes it does not.
The laptop also gets very hot if you are not sitting in A/C. Not sure if it is this specific laptop or it is a general issue
TBH, Google Assistant is not that much better. In the last few months it has become absurdly racist against my Italian accent, replying to me in Italian after I ask stuff in English - and getting the question wrong anyway.
It's ridiculous how poor in functionality all of them are... The best they can do is, what? Creating schedule entries, for me. All I ask Google is the weather, time, some search when I'm lazy and translate (the voice translate app itself is great btw). Feeling like a total corporate bitch saying "Hey Google" every time, too :D
This is supposed to be a personal assistant. And I have a whole list of what it could do for me, personally. But it doesn't.
I've been trying to figure out how to hook Google's speech recognition and voice into other apps, since they're great and it's 99% of what I need, hands-free control and feedback. Maybe they should make that easy, preferably offline and let other people create their own personal assistant modules or something.
Isn't half the point of these digital assistants that they do have the context?
I've got no idea about Siri, but the android one ties into your Google account to get your calendar and mail so it can get context about up coming travel etc.
> If the person asking the question lives in Ohio, they may actually be talking about London OH (or Dublin OH).
I would seriously doubt this assumption. Why on earth should someone living in a state specifically ask for the local time in a different location within that same state?
On the contrary, this context information would make it much more likely that the person actually meant "London, England". Except if there is a timezone border going through the state, of course.
However, I obviously agree with your general point regarding the severe limitations of what we currently call "AI" and how little "intelligence" there actually is.
> A simple "Which London?" could've removed the ambiguity and given Siri the opportunity to learn something about that particular person (that London, England is more important to him than London in Canada).
IMO I would be very disappointed if Siri started asking clarifying questions at a significantly higher rate. Siri is already a bit too chatty, and I never feel like having an extended conversation with her.
I’d rather she just say the wrong thing (but make it clear that the answer is for a specific London, e.g. “The time in London Ontario is...”) and I can correct her. It’s the same number of conversational “turns”, but in the happy path when she actually gets it right the first time, it’s one-shot and done.
It’s a lot harder to get signal on this for learning, but I feel like there are ways around this as well. (Maybe saying “thanks” can signal she got something right, and prefixing the next utterance with “no” could signal it was wrong...)
All of these situations are fairly complicated. So the best thing to do ( even for a human assistant ) is to ask more questions. That is the best way to get clarification rather than just trying to figure out what the context is. The assistant should be able to ask a simple question: "Are you talking about the London in UK or Ohio?"
The problem is not just that it is wrong, nor that it doesn't have enough personal information, but that it lacks proper personalisation and the ability to learn.
You can't reply with "no Siri, not that London" and have it remember. It doesn't learn your voice among the people who normally use your Siri in your household.
"Artificial intelligence" is always going to make mistakes, as do real humans. Humans can perform unsupervised learning - in fact it's one of the key skills that employers like to select on! Until AI can learn in context it's going to be very limited.
Fun etymology fact: Both Frankfurts get their name from a simple description of the same thing, which they both are instances of: settlements adjacent to a Frankish river ford (place where the river is shallow enough to cross without a bridge). The English "ford" and German "Furt" come from the same root; hence Francoford / Franken-Furt / Frankfurt. The "an der Oder" (at the Oder) and "am Main" (at the Main) suffixes are clarifications to describe which river is involved.
Exactly. And what about if I initially DO specify it's London Ontario that I'm interested in but five minutes later I ask again, referring to it simply as "London".. shouldn't a "better Siri" come back with 'assuming you still mean London Ontario, it's...' ?
I think it's possible that general acceptance of these non-AI gimmicks being referred to as "AI" will end up pushing genuine progress in true AI further into the future.
I would summarise it as: use separate keywords instead of sentences. "Change Light Bulb" instead of "how to change a light bulb". "Black Science Guy", "Kevin Durant height", "rails has_many api", etc...
Recently Google got much better in understanding full sentences and there are tons of SEO optimized pages for certain phrases. Nevertheless, using keywords is what I imagine advanced users do.
I wonder if anyone is working on special languages for talking to voice interfaces. Maybe a reduced grammar would allow for better recognition accuracy and reliability. And we could get more helpful corrections.
The problem reminds me of the difficulty of programming in applescript. In applescript, articles like "the" can be inserted optionally in the code, and there are lots of equivalent ways to write things, i.e. "if x equals y" is the same as "if x is equal to y". As a result I never remember the syntax, and error messages are less helpful.
I certainly prefer DDG here. There is no ambiguity in "11:00 EST to UTC". EST is always UTC - 5, EDT is always UTC - 4, and ET could mean either depending on the time of the year. Google isn't making an arbitrary decision to deal with ambiguity it's ignoring the specificity in the query and reinterpreting it (as ET rather than EST)
Just because it's hard to implement right doesen't mean Apple should get a pass. I mean, they aren't forced to create a voice assistant by regulation. If they can't make a good one, (and this applies to Amazon and Microsoft too) they should have just left it in the lab until they can.
But the post points out that Siri from other Apple devices gets it right. Apple’s “generic knowledge system” can answer this. It’s only Apple Watch which has trouble.
That’s kind of weird. It’s not that Siri is especially bad. It’s that “Siri” is something different depending on how you query it. Other online search systems aren’t like that, and integration and consistency are typically Apple’s forte.
> Where does such unpredictability come from? How can Siri transcribe the words correctly but fail to deliver the right answer?
Voice assistants generally use both the text transcription and a bunch of contextual metadata as input. That metadata could include things like what's currently visible on the screen, your location, your recent queries, etc.
So even though the underlying algorithms powering the assistant may be deterministic, the input data between two seemingly identical queries could vary quite a bit.
For instance, Siri almost certainly has context around the previous questions you've asked. It would be reasonable to assume that if an assistant received two identical questions back-to-back the initial answer was wrong.
In that scenario, the assistant might decide to use the a different answer (perhaps one that had a lower ranking) in an attempt to get it right.
Editorial Channel
What the content says
+0.20
Article 19Freedom of Expression
High Practice
Editorial
+0.20
SETL
+0.10
Author freely expresses sharp criticism of Apple's Siri design without apparent censorship or self-restraint. The post exemplifies unfettered freedom of opinion and expression toward a powerful corporation.
FW Ratio: 60%
Observable Facts
The author publishes direct criticism of Apple ('Stupid and slow is a heck of a combination') without apparent fear or self-censorship.
The site includes an RSS feed enabling free subscription and distribution of content.
No paywall, login requirement, or apparent access restriction visible on the domain.
Inferences
The author's willingness to publicly critique a major technology corporation demonstrates active practice of freedom of expression.
The site's unrestricted access model and RSS infrastructure structurally support the free distribution of opinion content.
ND
PreamblePreamble
No engagement with preamble themes (dignity, rights, freedom, justice).
ND
Article 1Freedom, Equality, Brotherhood
No engagement with equal dignity and rights.
ND
Article 2Non-Discrimination
No engagement with non-discrimination.
ND
Article 3Life, Liberty, Security
No engagement with right to life, liberty, security.
ND
Article 4No Slavery
No engagement with slavery.
ND
Article 5No Torture
No engagement with torture/cruel treatment.
ND
Article 6Legal Personhood
No engagement with recognition before law.
ND
Article 7Equality Before Law
No engagement with equal protection under law.
ND
Article 8Right to Remedy
No engagement with right to remedy.
ND
Article 9No Arbitrary Detention
No engagement with freedom from detention.
ND
Article 10Fair Hearing
No engagement with right to fair trial.
ND
Article 11Presumption of Innocence
No engagement with presumption of innocence.
ND
Article 12Privacy
No engagement with right to privacy.
ND
Article 13Freedom of Movement
No engagement with freedom of movement.
ND
Article 14Asylum
No engagement with right to asylum.
ND
Article 15Nationality
No engagement with right to nationality.
ND
Article 16Marriage & Family
No engagement with right to marry/family.
ND
Article 17Property
No engagement with right to property.
ND
Article 18Freedom of Thought
No engagement with freedom of conscience/religion.
ND
Article 20Assembly & Association
No engagement with freedom of association.
ND
Article 21Political Participation
No engagement with right to political participation.
ND
Article 22Social Security
No engagement with right to social security.
ND
Article 23Work & Equal Pay
No engagement with labor rights.
ND
Article 24Rest & Leisure
No engagement with right to rest/leisure.
ND
Article 25Standard of Living
No engagement with right to adequate standard of living.
ND
Article 26Education
No engagement with right to education.
ND
Article 27Cultural Participation
No engagement with cultural participation.
ND
Article 28Social & International Order
No engagement with just international order.
ND
Article 29Duties to Community
No engagement with duties to community.
ND
Article 30No Destruction of Rights
No engagement with freedom from right destruction.
Structural Channel
What the site does
+0.15
Article 19Freedom of Expression
High Practice
Structural
+0.15
Context Modifier
ND
SETL
+0.10
Site provides public RSS feed and unrestricted web access, supporting infrastructure for free distribution of diverse opinions. No paywall, login wall, or content gate barriers.
ND
PreamblePreamble
No structural signals related to preamble.
ND
Article 1Freedom, Equality, Brotherhood
No structural signals.
ND
Article 2Non-Discrimination
No structural signals.
ND
Article 3Life, Liberty, Security
No structural signals.
ND
Article 4No Slavery
No structural signals.
ND
Article 5No Torture
No structural signals.
ND
Article 6Legal Personhood
No structural signals.
ND
Article 7Equality Before Law
No structural signals.
ND
Article 8Right to Remedy
No structural signals.
ND
Article 9No Arbitrary Detention
No structural signals.
ND
Article 10Fair Hearing
No structural signals.
ND
Article 11Presumption of Innocence
No structural signals.
ND
Article 12Privacy
No engagement with privacy infrastructure.
ND
Article 13Freedom of Movement
No structural signals.
ND
Article 14Asylum
No structural signals.
ND
Article 15Nationality
No structural signals.
ND
Article 16Marriage & Family
No structural signals.
ND
Article 17Property
No structural signals.
ND
Article 18Freedom of Thought
No structural signals.
ND
Article 20Assembly & Association
No structural signals.
ND
Article 21Political Participation
No structural signals.
ND
Article 22Social Security
No structural signals.
ND
Article 23Work & Equal Pay
No structural signals.
ND
Article 24Rest & Leisure
No structural signals.
ND
Article 25Standard of Living
No structural signals.
ND
Article 26Education
No engagement with educational access/infrastructure.
ND
Article 27Cultural Participation
No structural signals.
ND
Article 28Social & International Order
No structural signals.
ND
Article 29Duties to Community
No structural signals.
ND
Article 30No Destruction of Rights
No structural signals.
Supplementary Signals
How this content communicates, beyond directional lean. Learn more
build 6157e1d+ai0o · deployed 2026-02-28 16:55 UTC · evaluated 2026-02-28 16:29:11 UTC
Support HN HRCB
Each evaluation uses real API credits. HN HRCB runs on donations — no ads, no paywalls.
If you find it useful, please consider helping keep it running.