Model Comparison
Model Editorial Structural Class Conf SETL Theme
claude-haiku-4-5-20251001 +0.20 +0.15 Mild positive 0.03 0.10 Free Expression
@cf/meta/llama-3.3-70b-instruct-fp8-fast lite 0.00 ND Neutral 0.90 0.00 No human rights theme
@cf/meta/llama-4-scout-17b-16e-instruct lite 0.00 ND Neutral 0.90 0.00 Technology
Section claude-haiku-4-5-20251001 @cf/meta/llama-3.3-70b-instruct-fp8-fast lite @cf/meta/llama-4-scout-17b-16e-instruct lite
Preamble ND ND ND
Article 1 ND ND ND
Article 2 ND ND ND
Article 3 ND ND ND
Article 4 ND ND ND
Article 5 ND ND ND
Article 6 ND ND ND
Article 7 ND ND ND
Article 8 ND ND ND
Article 9 ND ND ND
Article 10 ND ND ND
Article 11 ND ND ND
Article 12 ND ND ND
Article 13 ND ND ND
Article 14 ND ND ND
Article 15 ND ND ND
Article 16 ND ND ND
Article 17 ND ND ND
Article 18 ND ND ND
Article 19 0.18 ND ND
Article 20 ND ND ND
Article 21 ND ND ND
Article 22 ND ND ND
Article 23 ND ND ND
Article 24 ND ND ND
Article 25 ND ND ND
Article 26 ND ND ND
Article 27 ND ND ND
Article 28 ND ND ND
Article 29 ND ND ND
Article 30 ND ND ND
+0.20 Siri, What Time Is It in London? (daringfireball.net S:+0.15 )
728 points by jmsflknr 2108 days ago | 748 comments on HN | Mild positive Editorial · v3.7 · 2026-02-28 13:38:33
Summary Free Expression Acknowledges
Gruber's blog post criticizes Apple's Siri for inconsistent results when asked 'What time is it in London?', comparing it unfavorably to Google, DuckDuckGo, Alexa, and Bing. The post exemplifies freedom of expression (Article 19) through public, unrestricted criticism of a major corporation and features an accessible publication platform (RSS feed, no paywall). No other human rights provisions are engaged.
Article Heatmap
Preamble: ND — Preamble Preamble: No Data — Preamble P Article 1: ND — Freedom, Equality, Brotherhood Article 1: No Data — Freedom, Equality, Brotherhood 1 Article 2: ND — Non-Discrimination Article 2: No Data — Non-Discrimination 2 Article 3: ND — Life, Liberty, Security Article 3: No Data — Life, Liberty, Security 3 Article 4: ND — No Slavery Article 4: No Data — No Slavery 4 Article 5: ND — No Torture Article 5: No Data — No Torture 5 Article 6: ND — Legal Personhood Article 6: No Data — Legal Personhood 6 Article 7: ND — Equality Before Law Article 7: No Data — Equality Before Law 7 Article 8: ND — Right to Remedy Article 8: No Data — Right to Remedy 8 Article 9: ND — No Arbitrary Detention Article 9: No Data — No Arbitrary Detention 9 Article 10: ND — Fair Hearing Article 10: No Data — Fair Hearing 10 Article 11: ND — Presumption of Innocence Article 11: No Data — Presumption of Innocence 11 Article 12: ND — Privacy Article 12: No Data — Privacy 12 Article 13: ND — Freedom of Movement Article 13: No Data — Freedom of Movement 13 Article 14: ND — Asylum Article 14: No Data — Asylum 14 Article 15: ND — Nationality Article 15: No Data — Nationality 15 Article 16: ND — Marriage & Family Article 16: No Data — Marriage & Family 16 Article 17: ND — Property Article 17: No Data — Property 17 Article 18: ND — Freedom of Thought Article 18: No Data — Freedom of Thought 18 Article 19: +0.18 — Freedom of Expression 19 Article 20: ND — Assembly & Association Article 20: No Data — Assembly & Association 20 Article 21: ND — Political Participation Article 21: No Data — Political Participation 21 Article 22: ND — Social Security Article 22: No Data — Social Security 22 Article 23: ND — Work & Equal Pay Article 23: No Data — Work & Equal Pay 23 Article 24: ND — Rest & Leisure Article 24: No Data — Rest & Leisure 24 Article 25: ND — Standard of Living Article 25: No Data — Standard of Living 25 Article 26: ND — Education Article 26: No Data — Education 26 Article 27: ND — Cultural Participation Article 27: No Data — Cultural Participation 27 Article 28: ND — Social & International Order Article 28: No Data — Social & International Order 28 Article 29: ND — Duties to Community Article 29: No Data — Duties to Community 29 Article 30: ND — No Destruction of Rights Article 30: No Data — No Destruction of Rights 30
Negative Neutral Positive No Data
Aggregates
Editorial Mean +0.20 Structural Mean +0.15
Weighted Mean +0.18 Unweighted Mean +0.18
Max +0.18 Article 19 Min +0.18 Article 19
Signal 1 No Data 30
Volatility 0.00 (Low)
Negative 0 Channels E: 0.6 S: 0.4
SETL +0.10 Editorial-dominant
FW Ratio 60% 3 facts · 2 inferences
Evidence 3% coverage
1H 30 ND
Theme Radar
Foundation Security Legal Privacy & Movement Personal Expression Economic & Social Cultural Order & Duties Foundation: 0.00 (0 articles) Security: 0.00 (0 articles) Legal: 0.00 (0 articles) Privacy & Movement: 0.00 (0 articles) Personal: 0.00 (0 articles) Expression: 0.18 (1 articles) Economic & Social: 0.00 (0 articles) Cultural: 0.00 (0 articles) Order & Duties: 0.00 (0 articles)
HN Discussion 20 top-level · 30 replies
Hamuko 2020-05-22 10:08 UTC link
Same thing happens with Frankfurt. If I order something from Amazon.de, my package usually goes through Frankfurt. When I checked my package tracker app on my iPhone, I was surprised to find out that Frankfurt is actually on the Germany-Poland border.

Turns out that if I enter "Frankfurt, Germany" into Apple Maps (which I assume is what the package tracker app does), it takes me to "Frankfurt (Oder), Germany" instead of "Frankfurt am Main, Germany".

lqet 2020-05-22 10:14 UTC link
Ranking and NLP aren't easy. If you are asking a slightly related question (for example, "What is the weather in London"), and if you are living at some place where the nearest major town is called London, but is not London in England, you would expect it to give you the weather in "your" London. However, it you are asking for the time in a particular city, then the ranking should of course consider whether the timezone of the city you asked for is different than your own - it makes no sense to ask for the time in a city which lies in your timezone. Then again, if the distance from your location to that other London is greater than a certain threshold, the question could imply that you actually do not know whether the city lies in the same timezone as your location.

All these thresholds or ranking factors seem to come intuitively to humans (I would guess a good intuition for them is actually a sign of intelligence), but it seems to be incredibly hard to capture them in ranking.

As others have pointed out, a solution here would be to make Siri more conversational. A simple "Which London?" could've removed the ambiguity and given Siri the opportunity to learn something about that particular person (that London, England is more important to him than London in Canada).

jka 2020-05-22 10:21 UTC link
To throw a tongue-in-cheek additional example in here: the population of London, Ontario is (significantly) greater than the population of the City of London, UK.

Reality is hard. And with machine learning (especially proprietary, remotely-hosted machine learning) there's rarely a way to pinpoint a line of code and say: "this is what happened and why you're now frustrated and firing hypothetical personal assistants".

DangerousPie 2020-05-22 10:24 UTC link
It's ridiculous how Siri is still this shitty. I have an 11 Pro and even on such an expensive phone I can't really trust it to do anything more advanced than set timers. Every few months I try to do something else and just get annoyed at how bad it is.

Before lockdown I even had it disabled entirely because it would get activated randomly from time to time, even if nobody in the vicinity said anything remotely close to "Hey Siri".

dustinmoris 2020-05-22 10:28 UTC link
An actual assistant would have a lot of context. Do I know that you're going to travel to London, UK for a break next week and therefore I would naturally assume that you are interesting in the time there.

However, are you planning to visit your parents in London, Canada this weekend? Then an assistant who would still answer with the time of London, UK would maybe also not be the smartest?

So really context is everything and making broad statements that if an assistant was to answer with anything but London, UK should get fired is something that someone would say, who IMHO should get fired. shrug

Also, IMHO, if someone doesn't know that machines don't have human context and therefore doesn't know to ask their digital assistant "What is the time in London, UK" when they want to know the time in London, UK, then maybe they should get fired from their tech job. shrug

mojuba 2020-05-22 10:39 UTC link
This example illustrates how difficult AGI is and how far we are from it. We, humans, tend to take advantage of the context to make communication simpler and shorter. Just think about all the implications of this one simple question: what time is it in London? Or e.g. how can I get from London to Dublin?

If the person asking the question lives in Ohio, they may actually be talking about London OH (or Dublin OH). Some people in neighbouring states may mean the same, though they will be more likely to mention the state. However, how close should you be to London, OH even within the state to mean the Ohio one and not the UK one? How close is close enough? Is a few hours of driving close enough? A 3 hour flight? What if I'm roughly at 6 hours from London OH and 7 hours from London UK?

Further, if the person is a British expat in Ohio, especially if they are working for a multinational business (or not), they would more likely mean London UK. German expats, though? Russians? Or an Irish person who lives in Amsterdam having some relatives in Ohio US, looking to book a flight to Dublin. Etc. etc.

There are so many contextual layers here that even human assistants can occasionally get it wrong, and without the context the task becomes insurmountable for the "AI" algorithms. That is not to say virtual assistants are useless, just that selling them as "AI" is a big lie, bigger than even those who market these algorithms as "AI" think it is.

trombonechamp 2020-05-22 10:50 UTC link
For a related example, type "11:00 EST to UTC" into Google and DuckDuckGo. Google says 15:00 because it interprets "EST" colloquially as "the current time on east coast US". DuckDuckGo says 16:00 because it interprets EST literally as "the time on east coast when it isn't daylight savings time" (compared to ET or EDT). It isn't clear which behavior is more desirable.
fridek 2020-05-22 11:13 UTC link
I now have most of my adult years learning how to construct a phrase that gets the right results for Google, and later Assistant. It got to the point where I'm certain it must be a headache for whatever team is trying to support natural language processing in these - all proficient users ask for some artificial gibberish and get where they want to be.

Here comes my favourite brain freeze moment - recently my parents asked me to explain this to them. How do you construct a good search phrase? My brain blanked. I HAVE NO IDEA. It seems I have learned fluent Goonglish without noticing, and now can't explain the grammar or vocabulary of it.

brownbat 2020-05-22 11:21 UTC link
Aza Raskin's Ubiquity was such a clear model of how to build voice interfaces the right way, and it wasn't even a voice interface. It was a bit of a launcher that tied APIs together on the web.

Let users create and share small commands. Create a simple natural language for commands that are easy to program, extend, and remember, and narrow the scope of inputs the voice engine has to deal with.

It was so beautiful and effective and just light years ahead of what we're getting.

Microsoft gets a special mention for lost potential here. Their voice system in Windows could be a way to navigate the layered menus of the OS, but it is mostly focused on answering general queries. Voice is a great replacement for the program launcher, except it's not customizable, but that's about the extent of how much you can control the system with it. Let me do anything buried in the control panel, show me everything you know about a process when I ask, solve that first, then worry later about telling me how big the moon is. You make an OS, don't forget what that is.

hn_throwaway_99 2020-05-22 13:22 UTC link
TBH I always feel amazed about how worked up people get about stuff like this, especially people familiar with software who should know that there are millions, maybe billions of edge cases like this in a generic knowledge system, and thus at least it's easy to make a mistake like this. I mean, the time it took him to write his blog post is probably more than all the times it would take him to follow up with "What time is it in London, England?" It reminds me of someone who commented that "there must not be any black people who work at Apple" because it pronounced "Malcolm X Blvd" as Malcolm 10 Blvd.

I mean, if anything, just appreciate how amazing humans are at differentiating these ambiguities.

ubermonkey 2020-05-22 13:29 UTC link
Interesting. I can see how the basic algorithm might go "I've been asked for the time (or temp or whatever) in cityname. Citynames are routinely reused globally; what is the closest such cityname?"

But that fails completely when you get to names like London (or Paris or Moscow or Cairo).

But it happens with people, too. I'm from Mississippi, though I haven't lived there since I left for college. I now live in Houston. At a family reunion many years ago, I ran into a cousin I hadn't seen since we were kids. She asked where I was living, and I told her.

"Oh, isn't it terrible about that wreck?" she asked.

Baffled, I asked for more information. "Oh, you know, that wreck over on 406!"

I did not know. "I'm sorry, Houston's really huge. I don't know what wreck you mean."

"Oh, did you mean you live in Houston, TEXAS? I thought you meant Houston, MISSISSIPPI!"

I was, at the time, about 30. I grew up in that state, and lived there until I went to college. And until that moment, I had never even HEARD of Houston, Mississippi (a metropolis, it turns out, of about 3600 people in the misbegotten northeast corner of the state).

bgentry 2020-05-22 14:09 UTC link
The craziest and most confusing behavior of Siri for me is:

Sometimes you can ask a question and watch it be perfectly transcribed in real time, but then receive a nonsensical answer from it. Ask the exact same question immediately after on the same device, transcribed exactly the same way, and get the correct answer.

Where does such unpredictability come from? How can Siri transcribe the words correctly but fail to deliver the right answer?

jvolkman 2020-05-22 14:34 UTC link
At some point about a year ago I noticed that I could no longer ask my Google Home devices "what's the weather?". I'd just get a generic "I don't understand" response. But more specific queries such as "What's the weather in Seattle?" would work.

After a couple of weeks of this, I somehow got the idea that it was related to the devices' configured locations. And sure enough, telling the Home that I lived in the next city over fixed the problem.

So I started a binary search and eventually found that the issue was limited to my ~10x20 block Seattle neighborhood - basically the outline shown when I search for its name in Google Maps. I then also realized that it applied to weather queries on my phone as well, but since the phone uses GPS rather than a specific location setting, I could only reproduce the broken and working behaviors by crossing one of the neighborhood boundary streets.

Turns out it was some long-standing configuration issue with Knowledge Graph's entry for my neighborhood, and some recent code change in location-based weather queries began butting heads with it. Luckily I worked at Google at the time and was able to track down and pester people that could help fix the issue.

underbluewaters 2020-05-22 14:54 UTC link
"Hey Siri, set a timer for 5 minutes"

90% of the time works fine, and it's essentially all I use Siri for. It's very convenient when cooking and my hands are dirty. But 10% of the time I get something along the lines of...

"I'm sorry, but you don't have the Timer app installed".

"I'm sorry, but you don't have the Timer app installed".

"I'm sorry, but you don't have the Timer app installed".

"I'm sorry, but you don't have the Timer app installed".

It's infuriating because I know Siri is dumb so I use the same exact simple phrases to avoid confusion. Sometimes it works, sometimes it doesn't. It always transcribes the command accurately though! I've actually lost my temper and smashed an Apple Watch before over this. This is in my house, on a very reliable network, always with my phone within a reasonable distance.

kilo_bravo_3 2020-05-22 17:21 UTC link
Siri doesn't know that my front door is called "FRONT DOOR".

I only have one smart lock, which works perfectly, and it is called "FRONT DOOR" in HomeKit.

When I ask Siri about my FRONT DOOR she responds that she cannot find it.

When I ask Siri about the status of my DOOR, she responds with "The FRONT DOOR is locked/unlocked".

I'll then say 'Alright Siri you literally just used the phrase "FRONT DOOR" five seconds ago and the text transcript on the screen says "FRONT DOOR" hey Siri is my FRONT DOOR locked'

Siri: WTF are you talking about? You don't have a FRONT DOOR.

"Hey Siri is my door locked"

Siri: Your FRONT DOOR is locked.

Google and Alexa handle things flawlessly.

ngngngng 2020-05-22 17:30 UTC link
There's a Woodland Hills in Utah and a Woodland Hills in California. If you ask Google what the weather is in Woodland Hills, it will ALWAYS give you the weather for California. Even if your current location is Woodland Hills, Utah and even if your address is set to Woodland Hills Utah in your Google account or Google home.
killion 2020-05-22 17:45 UTC link
The fact that there isn't a feedback mechanism to let the Siri team know that it responded incorrectly tells me everything I need to know.

Until they have real metrics around how often Siri fails they will continue to think that their correct response rate is great.

dreamcompiler 2020-05-22 19:13 UTC link
Once I had a new assistant and I asked her to book me a flight to Boston. She went to the travel booking system, typed in Boston, then called me back confused. "Which Boston do you want? There are 8 of them?"

I was caught off guard and told her I'd prefer the one in Massachusetts. I did not fire her. She was young, had a poor general education, and had never traveled outside her home state. Those things do not make her stupid.

wpowiertowski 2020-05-22 23:15 UTC link
My usual dialog with siri:

Me: "Hey Siri, play Radiolab podcast"

Siri: Which Radiolab podcast, Radiolab or Radiolab: More Perfect"

Me: "Radiolab"

Siri: Which Radiolab podcast, Radiolab or Radiolab: More Perfect"

Me: "Radiolab"

Siri: Which Radiolab podcast, Radiolab or Radiolab: More Perfect"

...

Me: "The first one"

Siri: "I don't know >the first one<"

Me: "Siri you're useless"

Siri: "That's not nice"

Me: "Could be but it's true"

psaux 2020-05-23 05:30 UTC link
Might as well comment, first employee at Siri here. This result “maybe” should provide ambiguity resolution, but where does it stop. The she/he who compared it to Google was right on. Siri provides singular results in most cases vs multiple search style results. We did use geo for locality based results in the past. This would solve the problem the OP mentioned, not sure if they call location for these requests now. The other person who mentioned we can’t/couldn’t train on data is correct too. Again, privacy first. Be proud, Apple cares a lot. When the Siri commercials hit (No one told us there would be commercials) when we launched, we got decimated, and couldn’t debug the issues, user utterances were not allowed to be logged. Luckily, after much sleep deprivation, one of my engineers (love you Stu) said, “hey, aren’t they running commercials” to all our surprise. We convinced the privacy team to let us log word parts. Then we started to see words that were present in the commercials. Fun fact, also happened when Tom Cruise was presenting at the Academy Awards. We had millions+ asks all at the same time, again word parts. “height”, “tom”, “foot”, etc.
nivenkos 2020-05-22 10:21 UTC link
I remember Deutsche Bahn's website used to do the same with Freiburg and Freiburg im Breisgau. Almost made that mistake once.
gruturo 2020-05-22 10:22 UTC link
My company (based in Frankfurt am Main, Germany) had an online business travel reservation system which helpfully pre-filled the starting point of all new trips to Frankfurt, Kentucky (which is not even the right spelling of the city).

I filed a request to NOT pre-fill the starting city to a place in Kentucky and got a polite but firm reply that this was the default list provided by the 3rd party online booking engine and cannot be customized. Sigh.

CamperBob2 2020-05-22 10:24 UTC link
As others have pointed out, a solution here would be to make Siri more conversational.

But that would make it almost as smart as an Infocom game from 1981. Something, something, doesn't scale, mumble, something...

bnt 2020-05-22 10:27 UTC link
I’ve had Siri disabled for years. Even the basic “call home” works every 3rd time. I try it for a few minutes with every new iOS update only to see it’s still the same dumbster fire.
amelius 2020-05-22 10:27 UTC link
The problem sounds a little like collaborative filtering. If you have a certain affinity with cities A,B,C, then you can compute the expected affinity with a city X by looking at the affinities other people have with X, and their affinities with A,B,C.

Instead of looking at people, you can also scrape websites to get the relations. But here you may get a recursive problem because if a website speaks of "London", you might not know in advance which London they speak of.

Veen 2020-05-22 10:28 UTC link
Yes, but in the same way people don't usually mean "London, Ontario" when they say London, they also don't usually mean the City of London (which, for the benefit of people who may not know, is a tiny portion of London with a population less than 10,000).
fireattack 2020-05-22 10:32 UTC link
But if none of these contexts you mentioned exist (which is the case here), what would an actual assistant answer?
nixpulvis 2020-05-22 10:34 UTC link
ha! I was wondering how hard it'd be to find someone making an appeal to "context". While you're not exactly wrong, you are. Parsing meaning is something humans are surprisingly good at, and trained for.

I think the most damning part is how, at the bottom, he list a handful of other "smart" assistants which correctly list London UK's time... for now.

But, his point about consistency and slowness is exactly why I never use these shitty voice assistants. If I'm going to be interacting with some pedantic robot, I generally want to be able to edit the text of my request.

TLDR; voice assistants suck.

qayxc 2020-05-22 10:34 UTC link
In the absence of context, the best answer is London, UK.

You actually need to imagine additional context to make any other answer plausible.

If you hear the sound of hooves clip-clopping nearby, you think horses, not zebras.

MachineGunPablo 2020-05-22 10:40 UTC link
Ouch
brnt 2020-05-22 10:40 UTC link
If you ever cross the border from Frankfurt/Oder, the first thing you see in Poland is a giant Amazon building. Despite that building being there, Polish people still can't order at Amazon.
mrshu 2020-05-22 10:43 UTC link
A similar thing happened with some football fans that made it to "Frankfurt (Oder)" instead of "Frankfurt am Main" for the Europa League's semifinals.

Judging from the published pictures [0], it looks like Google Maps may lead to the same result as Apple Maps for "Frankfurt".

[0] https://talksport.com/football/529808/benfica-fan-wrong-fran...

chappar 2020-05-22 10:44 UTC link
I had never used an Apple product before the company which I joined recently gave me a MacBook Pro. I am really surprised how bad the product quality is. The calendar notification is very random. Sometimes it fires, sometimes it does not. I have missed couple of meetings because notification popped after the meeting was over. Similarly the keyboard shortcut is random. Sometimes it opens the app, sometimes it does not. The laptop also gets very hot if you are not sitting in A/C. Not sure if it is this specific laptop or it is a general issue
toyg 2020-05-22 10:47 UTC link
TBH, Google Assistant is not that much better. In the last few months it has become absurdly racist against my Italian accent, replying to me in Italian after I ask stuff in English - and getting the question wrong anyway.

But yes, Siri is the worst.

jotm 2020-05-22 10:49 UTC link
It's ridiculous how poor in functionality all of them are... The best they can do is, what? Creating schedule entries, for me. All I ask Google is the weather, time, some search when I'm lazy and translate (the voice translate app itself is great btw). Feeling like a total corporate bitch saying "Hey Google" every time, too :D

This is supposed to be a personal assistant. And I have a whole list of what it could do for me, personally. But it doesn't.

I've been trying to figure out how to hook Google's speech recognition and voice into other apps, since they're great and it's 99% of what I need, hands-free control and feedback. Maybe they should make that easy, preferably offline and let other people create their own personal assistant modules or something.

mattlondon 2020-05-22 10:51 UTC link
Isn't half the point of these digital assistants that they do have the context?

I've got no idea about Siri, but the android one ties into your Google account to get your calendar and mail so it can get context about up coming travel etc.

Slartie 2020-05-22 10:55 UTC link
> If the person asking the question lives in Ohio, they may actually be talking about London OH (or Dublin OH).

I would seriously doubt this assumption. Why on earth should someone living in a state specifically ask for the local time in a different location within that same state?

On the contrary, this context information would make it much more likely that the person actually meant "London, England". Except if there is a timezone border going through the state, of course.

However, I obviously agree with your general point regarding the severe limitations of what we currently call "AI" and how little "intelligence" there actually is.

ninkendo 2020-05-22 10:57 UTC link
> A simple "Which London?" could've removed the ambiguity and given Siri the opportunity to learn something about that particular person (that London, England is more important to him than London in Canada).

IMO I would be very disappointed if Siri started asking clarifying questions at a significantly higher rate. Siri is already a bit too chatty, and I never feel like having an extended conversation with her.

I’d rather she just say the wrong thing (but make it clear that the answer is for a specific London, e.g. “The time in London Ontario is...”) and I can correct her. It’s the same number of conversational “turns”, but in the happy path when she actually gets it right the first time, it’s one-shot and done.

It’s a lot harder to get signal on this for learning, but I feel like there are ways around this as well. (Maybe saying “thanks” can signal she got something right, and prefixing the next utterance with “no” could signal it was wrong...)

t-shaped 2020-05-22 10:59 UTC link
All of these situations are fairly complicated. So the best thing to do ( even for a human assistant ) is to ask more questions. That is the best way to get clarification rather than just trying to figure out what the context is. The assistant should be able to ask a simple question: "Are you talking about the London in UK or Ohio?"
pjc50 2020-05-22 11:08 UTC link
The problem is not just that it is wrong, nor that it doesn't have enough personal information, but that it lacks proper personalisation and the ability to learn.

You can't reply with "no Siri, not that London" and have it remember. It doesn't learn your voice among the people who normally use your Siri in your household.

"Artificial intelligence" is always going to make mistakes, as do real humans. Humans can perform unsupervised learning - in fact it's one of the key skills that employers like to select on! Until AI can learn in context it's going to be very limited.

an_ko 2020-05-22 11:08 UTC link
Fun etymology fact: Both Frankfurts get their name from a simple description of the same thing, which they both are instances of: settlements adjacent to a Frankish river ford (place where the river is shallow enough to cross without a bridge). The English "ford" and German "Furt" come from the same root; hence Francoford / Franken-Furt / Frankfurt. The "an der Oder" (at the Oder) and "am Main" (at the Main) suffixes are clarifications to describe which river is involved.
BruceEel 2020-05-22 11:32 UTC link
Exactly. And what about if I initially DO specify it's London Ontario that I'm interested in but five minutes later I ask again, referring to it simply as "London".. shouldn't a "better Siri" come back with 'assuming you still mean London Ontario, it's...' ?

I think it's possible that general acceptance of these non-AI gimmicks being referred to as "AI" will end up pushing genuine progress in true AI further into the future.

calineczka 2020-05-22 11:45 UTC link
I would summarise it as: use separate keywords instead of sentences. "Change Light Bulb" instead of "how to change a light bulb". "Black Science Guy", "Kevin Durant height", "rails has_many api", etc...

Recently Google got much better in understanding full sentences and there are tons of SEO optimized pages for certain phrases. Nevertheless, using keywords is what I imagine advanced users do.

maliker 2020-05-22 12:14 UTC link
I wonder if anyone is working on special languages for talking to voice interfaces. Maybe a reduced grammar would allow for better recognition accuracy and reliability. And we could get more helpful corrections.

The problem reminds me of the difficulty of programming in applescript. In applescript, articles like "the" can be inserted optionally in the code, and there are lots of equivalent ways to write things, i.e. "if x equals y" is the same as "if x is equal to y". As a result I never remember the syntax, and error messages are less helpful.

polote 2020-05-22 12:16 UTC link
I have worked on that problem recently, and yes it is really hard.

You will be surprised on how many cities are named 'San francisco' in the world : https://en.wikipedia.org/wiki/San_Francisco_(disambiguation)

a1a1a1a1a1a1 2020-05-22 13:23 UTC link
I certainly prefer DDG here. There is no ambiguity in "11:00 EST to UTC". EST is always UTC - 5, EDT is always UTC - 4, and ET could mean either depending on the time of the year. Google isn't making an arbitrary decision to deal with ambiguity it's ignoring the specificity in the query and reinterpreting it (as ET rather than EST)
recursive 2020-05-22 13:30 UTC link
Just because it's hard to implement right doesen't mean Apple should get a pass. I mean, they aren't forced to create a voice assistant by regulation. If they can't make a good one, (and this applies to Amazon and Microsoft too) they should have just left it in the lab until they can.
ken 2020-05-22 14:08 UTC link
But the post points out that Siri from other Apple devices gets it right. Apple’s “generic knowledge system” can answer this. It’s only Apple Watch which has trouble.

That’s kind of weird. It’s not that Siri is especially bad. It’s that “Siri” is something different depending on how you query it. Other online search systems aren’t like that, and integration and consistency are typically Apple’s forte.

emiliobumachar 2020-05-22 14:24 UTC link
Siri may be a marvel of modern technology. If the competition does better, than it's reasonable to complain about Siri.
objclxt 2020-05-22 14:36 UTC link
> Where does such unpredictability come from? How can Siri transcribe the words correctly but fail to deliver the right answer?

Voice assistants generally use both the text transcription and a bunch of contextual metadata as input. That metadata could include things like what's currently visible on the screen, your location, your recent queries, etc.

So even though the underlying algorithms powering the assistant may be deterministic, the input data between two seemingly identical queries could vary quite a bit.

For instance, Siri almost certainly has context around the previous questions you've asked. It would be reasonable to assume that if an assistant received two identical questions back-to-back the initial answer was wrong.

In that scenario, the assistant might decide to use the a different answer (perhaps one that had a lower ranking) in an attempt to get it right.

Editorial Channel
What the content says
+0.20
Article 19 Freedom of Expression
High Practice
Editorial
+0.20
SETL
+0.10

Author freely expresses sharp criticism of Apple's Siri design without apparent censorship or self-restraint. The post exemplifies unfettered freedom of opinion and expression toward a powerful corporation.

ND
Preamble Preamble

No engagement with preamble themes (dignity, rights, freedom, justice).

ND
Article 1 Freedom, Equality, Brotherhood

No engagement with equal dignity and rights.

ND
Article 2 Non-Discrimination

No engagement with non-discrimination.

ND
Article 3 Life, Liberty, Security

No engagement with right to life, liberty, security.

ND
Article 4 No Slavery

No engagement with slavery.

ND
Article 5 No Torture

No engagement with torture/cruel treatment.

ND
Article 6 Legal Personhood

No engagement with recognition before law.

ND
Article 7 Equality Before Law

No engagement with equal protection under law.

ND
Article 8 Right to Remedy

No engagement with right to remedy.

ND
Article 9 No Arbitrary Detention

No engagement with freedom from detention.

ND
Article 10 Fair Hearing

No engagement with right to fair trial.

ND
Article 11 Presumption of Innocence

No engagement with presumption of innocence.

ND
Article 12 Privacy

No engagement with right to privacy.

ND
Article 13 Freedom of Movement

No engagement with freedom of movement.

ND
Article 14 Asylum

No engagement with right to asylum.

ND
Article 15 Nationality

No engagement with right to nationality.

ND
Article 16 Marriage & Family

No engagement with right to marry/family.

ND
Article 17 Property

No engagement with right to property.

ND
Article 18 Freedom of Thought

No engagement with freedom of conscience/religion.

ND
Article 20 Assembly & Association

No engagement with freedom of association.

ND
Article 21 Political Participation

No engagement with right to political participation.

ND
Article 22 Social Security

No engagement with right to social security.

ND
Article 23 Work & Equal Pay

No engagement with labor rights.

ND
Article 24 Rest & Leisure

No engagement with right to rest/leisure.

ND
Article 25 Standard of Living

No engagement with right to adequate standard of living.

ND
Article 26 Education

No engagement with right to education.

ND
Article 27 Cultural Participation

No engagement with cultural participation.

ND
Article 28 Social & International Order

No engagement with just international order.

ND
Article 29 Duties to Community

No engagement with duties to community.

ND
Article 30 No Destruction of Rights

No engagement with freedom from right destruction.

Structural Channel
What the site does
+0.15
Article 19 Freedom of Expression
High Practice
Structural
+0.15
Context Modifier
ND
SETL
+0.10

Site provides public RSS feed and unrestricted web access, supporting infrastructure for free distribution of diverse opinions. No paywall, login wall, or content gate barriers.

ND
Preamble Preamble

No structural signals related to preamble.

ND
Article 1 Freedom, Equality, Brotherhood

No structural signals.

ND
Article 2 Non-Discrimination

No structural signals.

ND
Article 3 Life, Liberty, Security

No structural signals.

ND
Article 4 No Slavery

No structural signals.

ND
Article 5 No Torture

No structural signals.

ND
Article 6 Legal Personhood

No structural signals.

ND
Article 7 Equality Before Law

No structural signals.

ND
Article 8 Right to Remedy

No structural signals.

ND
Article 9 No Arbitrary Detention

No structural signals.

ND
Article 10 Fair Hearing

No structural signals.

ND
Article 11 Presumption of Innocence

No structural signals.

ND
Article 12 Privacy

No engagement with privacy infrastructure.

ND
Article 13 Freedom of Movement

No structural signals.

ND
Article 14 Asylum

No structural signals.

ND
Article 15 Nationality

No structural signals.

ND
Article 16 Marriage & Family

No structural signals.

ND
Article 17 Property

No structural signals.

ND
Article 18 Freedom of Thought

No structural signals.

ND
Article 20 Assembly & Association

No structural signals.

ND
Article 21 Political Participation

No structural signals.

ND
Article 22 Social Security

No structural signals.

ND
Article 23 Work & Equal Pay

No structural signals.

ND
Article 24 Rest & Leisure

No structural signals.

ND
Article 25 Standard of Living

No structural signals.

ND
Article 26 Education

No engagement with educational access/infrastructure.

ND
Article 27 Cultural Participation

No structural signals.

ND
Article 28 Social & International Order

No structural signals.

ND
Article 29 Duties to Community

No structural signals.

ND
Article 30 No Destruction of Rights

No structural signals.

Supplementary Signals
How this content communicates, beyond directional lean. Learn more
Epistemic Quality
How well-sourced and evidence-based is this content?
0.54 medium claims
Sources
0.5
Evidence
0.6
Uncertainty
0.5
Purpose
0.6
Propaganda Flags
3 manipulative rhetoric techniques found
3 techniques detected
loaded language
"Stupid and slow is a heck of a combination" — harsh emotional language applied to Siri's performance.
exaggeration
Extrapolates one wrong answer to 'a serious cognitive deficiency that permeates everything they try to do.'
appeal to authority
Cites that DuckDuckGo, Google, Alexa, and Bing all got the answer correct, implying superior design.
Emotional Tone
Emotional character: positive/negative, intensity, authority
confrontational
Valence
-0.5
Arousal
0.6
Dominance
0.7
Transparency
Does the content identify its author and disclose interests?
0.67
✓ Author ✗ Conflicts ✓ Funding
More signals: context, framing & audience
Solution Orientation
Does this content offer solutions or only describe problems?
0.06 problem only
Reader Agency
0.1
Stakeholder Voice
Whose perspectives are represented in this content?
0.25 2 perspectives
Speaks: individual
About: corporation
Temporal Framing
Is this content looking backward, at the present, or forward?
present immediate
Geographic Scope
What geographic area does this content cover?
national
Canada, United States, United Kingdom
Complexity
How accessible is this content to a general audience?
accessible low jargon general
Audit Trail 7 entries
2026-02-28 13:38 eval Evaluated by claude-haiku-4-5-20251001: +0.18 (Mild positive)
2026-02-28 12:11 eval_success Lite evaluated: Neutral (0.00) - -
2026-02-28 12:11 rater_validation_warn Lite validation warnings for model llama-3.3-70b-wai: 0W 1R - -
2026-02-28 12:11 eval Evaluated by llama-3.3-70b-wai: 0.00 (Neutral)
2026-02-28 12:06 eval_success Lite evaluated: Neutral (0.00) - -
2026-02-28 12:06 eval Evaluated by llama-4-scout-wai: 0.00 (Neutral)
2026-02-28 12:06 rater_validation_warn Lite validation warnings for model llama-4-scout-wai: 0W 1R - -