Dream it. Build it. Grow it. Sign up now and you'll be up and running on DigitalOcean in just minutes.
Paul Kafasis engages in some excellent, self-inflicted nerd-snipping on One Foot Tsunami:
I asked my iPhone who won Super Bowls 1 through 60 (that’s “I” through “LX” in Super Bowl styling) and captured a screenshot of each result.
The results are utterly appalling:
So, how did Siri do? With the absolute most charitable interpretation, Siri correctly provided the winner of just 20 of the 58 Super Bowls that have been played. That’s an absolutely abysmal 34% completion percentage. If Siri were a quarterback, it would be drummed out of the NFL.
Some of the results are especially awful. For example, to the question “Who won Super Bowl XXIII?”, Siri responds with the number of times Bill Belichick has won or appeared in the Super Bowl—completely irrelevant.
John Gruber at Daring Fireball wrote a brutally (but fairly) titled follow-up, Siri Is Super Dumb and Getting Dumber, sharing the appalling results to his own query, “Who won the 2004 North Dakota high school boys’ state basketball championship?”
New Siri — powered by Apple Intelligence™ with ChatGPT integration enabled — gets the answer completely but plausibly wrong, which is the worst way to get it wrong. It’s also inconsistently wrong — I tried the same question four times, and got a different answer, all of them wrong, each time. It’s a complete failure.
We’ve all had the Siri experience of getting a clearly wrong or patently useless answer to our query. It’s gotten to the point where I merely roll my eyes and move on—I rarely even screenshot mistakes anymore.
But I do feel sorry for the Siri team. I have some good friends who work there, and I had occasion to work with the team on Siri responses a few years back. I know they cringe every time these failures hit the blogs. They know more than anyone just how much Siri needs to improve.
The latest scuttlebutt (from Mark Gurman at Bloomberg) is that longtime Apple exec Kim Vorrath is moving to Apple Intelligence in an effort to whip it into shape. I’ve watched Vorrath and her Program Office teams operate from the inside for many years. The biggest impact she and her team had across engineering was instilling discipline: every feature or bug fix had to be approved; tied to a specific release; and built, tested, and submitted on time. It was (is!) a time-intensive process—and engineering often complained about it, sometimes vocally—but the end result was a more defined, less kitchen-sink release each year. To a significant extent, her team is the reason why a feature may get announced at WWDC but not get released until the following spring. She provided engineering risk management.
I hope the Vorrath and the Siri team can make this work. I need them to make this work. The future promised by Apple Intelligence is too compelling for it to fail.
Graham Fraser, writing about the BBC, on BBC:
The BBC has complained to Apple after the tech giant's new iPhone feature generated a false headline about a high-profile murder in the United States.
Apple Intelligence, launched in the UK earlier this week, uses artificial intelligence (AI) to summarize and group together notifications.
Apple Intelligence is new to the U.K, but those of us in the U.S. have been ridiculing it for a month now. As John McClane said, “Welcome to the party, pal!”
This week, the AI-powered summary falsely made it appear BBC News had published an article claiming Luigi Mangione, the man arrested following the murder of healthcare insurance CEO Brian Thompson in New York, had shot himself. He has not.
Headlines are an editorial decision, and represent the voice of the publication. A poor summary can be embarrassing. A misleading one—as this was—can sully the publication.
"BBC News is the most trusted news media in the world," the BBC spokesperson added.
"It is essential to us that our audiences can trust any information or journalism published in our name and that includes notifications."
Apple can’t afford this bad press if Apple Intelligence is going to be taken seriously and drive hardware sales.
If Apple can’t address this quickly, they may have another egg freckles situation on their hands.[1]
To summarize: The handwriting recognition on the Apple Newton would fail, often in spectacular ways. Garry Trudeau “mocked the Newton in a weeklong arc of his comic strip Doonesbury, portraying it as a costly toy that served the same function as a cheap notepad, and using its accuracy problems to humorous effect. In one panel, Michael Doonesbury's Newton misreads the words "Catching on?" as "Egg Freckles", a phrase that became widely repeated as symbolic of the Newton's problems.” ↩︎