By John Gruber
WorkOS: Scalable, secure authentication, trusted by OpenAI, Cursor, Perplexity, and Vercel.
New (well, newish) Mac app from John Siracusa:
Hyperspace searches for files with identical contents within one or more folders. If it finds any, it can then reclaim the disk space taken by all but one of the identical files — without removing any of the files!
You can learn more about how this is done, if you’re interested, but the short version is that Hyperspace uses a standard feature of the macOS file system: space-saving clones. The Finder does the same thing when you duplicate a file.
I love everything about this app. I love the name — it “works” in like at least three ways. I love that it’s right up Siracusa’s alley. I love that Siracusa has talked about it, at wonderful length, on ATP and expounded upon it on his blog. I love that the premise sounds a little crazy but the explanation makes all the sense in the world. I love that this small, laser-focused utility is fully and splendidly documented. I love the way it looks. It’s got a great icon. I mean of course it would, but still, let’s celebrate how fun this is.
Not new, but new to me, is this delightful 7-minute short with a behind-the-scenes look at SNL’s cue card team, led by longtime main cue card guy Wally Feresten. Sometimes you just can’t beat analog.
As a kid I loved Richard Scarry’s books. As an adult I loved (and love) Chris Ware’s graphic novels. As a parent I loved reading Scarry’s books, again, with my son. So of course this essay from Ware, commemorating the 50th anniversary edition of Scarry’s Cars and Trucks and Things That Go, hit hard for me. Bet it will for you too.
Om Malik:
I have my own explanation, something my readers are familiar with, and it is the most obvious one. Just as Google is trapped in the 10-blue-link prison, which prevents it from doing something radical, Apple has its own golden handcuffs. It’s a company weighed down by its market capitalization and what stock market expects from it.
They lack the moral authority of Steve Jobs to defy the markets, streamline their product lineup, and focus the company. Instead, they do what a complex business often does: they do more. Could they have done a better job with iPadOS? Should Vision Pro receive more attention?
The answer to all those is yes. Apple has become a complex entity that can’t seem to ever have enough resources to provide the real Apple experience. What you get is “good enough.” And most of the time, I think it is enough — because what others have on the market is worse. They know how to build great hardware; it’s the software where they falter. In the case of Apple Intelligence, they have been caught short because others’ AI products, even when flawed, are significantly better than Apple’s own offerings.
Hardware inherently keeps a company honest in a way that software doesn’t. Hardware either works or it doesn’t. The only way to “upgrade” hardware is via installing newer software, or by taking the hardware apart and replacing physical components. It’s hard to think of a company, in any field, whose software is “better” than its hardware. Maybe Nintendo? But even with Nintendo, I’d say it’s more like their software is as good as their hardware. Also, an interesting thought that popped into my head reading Malik’s post just now: part of what makes Vision Pro so fascinating is that the software is better than the hardware. The hardware for immersive VR is so early-days that even the industry state-of-the-art — which is Vision Pro — stinks compared to where it’s going to be in even just five years. The 1984 Macintosh was a shitty computer with a 9-inch one-bit display, no hard drive, and an absurdly meager 128 kilobytes of RAM. But the software was amazing!
But the bigger, better point Malik makes is that “good enough” is enough to make Apple’s software seem ahead of its competition. I tried to make this point all the way back in 2007 with “Apple Needs a Nikon”, and I think the problem is worse now than it was then. No other company is even vaguely in Apple’s league. But Apple is sliding toward mediocrity on the software side. It very open for debate how far they’ve slipped. I, for one, would argue that they haven’t slipped far, and with an honest reckoning — especially with regard to everything related to Siri and AI — they can nip this in the bud. You might argue that they’ve slipped tremendously across the board. But what I don’t think is arguable is that their competition remains below Apple’s league. That’s what gives credence to the voices in Cupertino who are arguing that everything’s fine. Apple’s the only team in the top tier for UI design.
The best thing that could happen to Apple would be for Google to ship an Android Pixel experience that actually makes iPhone owners insanely jealous. Google is incapable of doing that through UI design. They’re incapable of catching up to Apple on hardware. But maybe on the AI front they can do it. Apple needs a rival.
Tesla’s share price has been having a hard time of it lately. The stock has lost about half its value since its all-time high back in December, and, since Musk took office alongside Donald Trump in January, dropped for 7 consecutive weeks, rebounding only ever-so-slightly last week, after Musk got the president of the United States to turn the White House lawn into a cheesy Tesla (sorry, Tesler) dealership. Tesla stock dropped another 5 percent today, on a day when the overall market was slightly up.
I bookmarked this Bryce Elder column at the Financial Times back on January 31, and now seems like a good time to link to it:
The usual explanation for when Tesla trading resembles a Pump.fun shitcoin is: “because Elon talks a lot”. Here’s JPMorgan analyst Ryan Brinkman to expand on the theme:
It’s not clear to us why Tesla shares traded as much as +5% higher in the aftermarket Wednesday, although we have some leading theories. Perhaps it was management’s statement that it had identified an achievable path to becoming worth more than the world’s five most valuable companies taken together (i.e., more than the $14.8 trillion combined market capitalizations of Apple, Microsoft, NVIDIA, Amazon, & Alphabet). Or maybe it was management’s belief that just one of its products has by itself the potential to generate “north of $10 trillion in revenue”. It may have even related to management guidance for 2026 (no financial targets were provided, but it was said to be “epic”) and for 2027 and 2028 (“ridiculously good”).
Brinkman, who has a long-standing “underweight” rating on Tesla, is beginning to sound a bit exasperated:
[T]he company’s financial performance and Bloomberg consensus for revenue, margin, earnings, and cash flow all keep coming down, but analyst price targets and the company’s share price keep going up. For instance, Tesla has missed Bloomberg consensus EBIT in 9 of the past 10 quarters by an average of -16.3%.
Consistently missing estimates is one thing. What Tesla has been doing is consistently missing lowered estimates. [...]
Tesla’s biggest asset is hyperbole. The more extreme the hyperbole, the more valuable it gets. Maybe after-hours market participants understand the dynamics better than Tesla bears, so are primed to park fundamentals and trade on vibes. Or maybe something else entirely is going on.
Sounds a lot like the other guy at the White House Auto Mall.
I’ve been commenting and expanding upon some of the commentary my piece prompted, and I have a few more coming, but it’s good to have Tsai collect a comprehensive overview.
Ray Maker, writing at DC Rainmaker:
This would not only be the first time Apple has created a non-watch heart rate sensor, but even more notably, the first time the company has enabled heart rate broadcasting over existing Bluetooth heart rate standards.
The question then becomes: Is it accurate?
Unfortunately, it turns out, that was not the question I should have started with. The real question to start with is: Is the heart rate function (accuracy aside), even usable? A lot of hours later, I have answers to both of those questions. And trust me, it’s a very mixed bag.
The answer:
It’s clear that any movement (even on a stationary bike) quickly leads to either dropouts or inaccurate heart rate. And outdoors running, it’s even worse. Ultimately, I don’t see any value in the heart rate sensor in this product, because it’s simply not good enough to be useful, even for casual use.
So maybe this feature is not soon coming to AirPods? I think there’s a good argument to be made that these are better than no heart rate monitor at all but also not nearly as good as an Apple Watch or dedicated device.
I’m a month late linking to it, but Chance Miller wrote a terrific review for 9to5Mac:
The last several releases from Beats, such as the Studio Buds Plus and Solo 4 headphones, have been powered by a custom Beats chip rather than an Apple-designed chip like what’s used in AirPods. For Beats, this has enabled better cross-platform support for Android users, but it’s also come at the cost of several popular features for Apple fans. For example, the Studio Buds Plus lack support for automatic in-ear detection, iCloud pairing, automatic device switching, personalized spatial audio, and more.
With the Powerbeats Pro 2, Beats has gone back to its roots and opted for an Apple-designed chip. The Powerbeats Pro 2 are powered by Apple’s H2 chip, the same chip used by the latest-generation AirPods Pro 2 and AirPods 4. This means you get the full suite of Apple-focused audio features.
The degree of shared engineering between Apple’s teams and Beats’s has always seemed odd to me. Sometimes it seems like Beats really is an independent subsidiary, focused on cross-platform headphones, and other times it feels like they’re making Apple products under a different brand label. The sweet spot seems to be about where they landed with these Powerbeats 2.
All of the aforementioned features and improvements make Powerbeats Pro 2 an incredibly compelling product, but Beats has one more thing: Powerbeats Pro 2 feature built-in heart rate monitoring.
Each Powerbeats Pro 2 earbud has a built-in heart rate monitor comprised of four components. First, there’s an LED sensor that emits green LED light at a rate of over 100 pulses per second. This light is emitted through the skin and hits your red blood cells. The photodiode then receives the reflected light from the red blood cells that is modulated by the red blood flow. There’s an optical lens that helps direct and separate the transmitted and received light, along with an accelerometer to ensure accuracy and consistency in data collection.
Beats adds that the Powerbeats Pro 2’s heart rate sensor technology is derived from Apple’s work on the Apple Watch.
It’s weird, but cool, that Beats has delivered in-ear heartbeat monitoring before Apple’s own AirPods have. But now it seems like a lock that this will be a feature in AirPods Pro 3, right?
What I always want in a review I read — and what I try to provide to readers through my own reviews — is a sense of whether a product is for me. Powerbeats Pro 2 aren’t for me — and I know it, because Miller’s review describes them so well. But they seem like a terrific product that a lot of people would prefer to AirPods Pro.
Sebastiaan de With, on X, linking to my “Something Is Rotten” piece last week:
Ex-MobileMe team here. This was a brutal time.
It was so bad that when he presented iCloud onstage, Steve said “I know what you’re thinking: why should I trust them? They’re the ones who gave us MobileMe!”
Michael Gartenberg (who worked at Apple in product marketing for a few years at the tail end of the Jobs era), responded (across two tweets):
When I was at Apple and Apple University was still around there was a whole course on MobileMe and how it was possible that things ended up the way they did. Fascinating to hear all the backstory.
One of the lessons of the Apple University course was much of the MobileMe debacle was directly because Jobs didn’t care about it. He was too preoccupied with the newest iPhone at the time. He didn’t even introduce the product, a lot of the stuff crossed his desk that he ignored.
Twitter-like social posts enforce brevity, but I suspect Gartenberg would agree that it wasn’t that Jobs didn’t care about MobileMe at all. It was that he didn’t think he had to care enough to devote his personal attention to it. Yes, Apple should offer web-based functionality for some online fundamentals (email, calendar, contacts...) and, more importantly, Apple should provide over-the-air Internet sync for that data between customers’ devices. And it should just work, in the way that a hard drive “just works” without Steve Jobs paying close attention to the current state of Apple’s file system team. But then it turned out MobileMe didn’t “just work”, and Jobs decided that he needed to pay laser-focused attention to starting over and building what we now know as iCloud (which is really quite good, very reliable, and I’d say long ago surpassed the “it just works” threshold). Steve Jobs’s final keynote — at WWDC 2011 — was largely focused on the announcement of iCloud.
Who’s got that role inside Apple today — someone with high standards, good taste, and clout within the company — for Siri and Apple Intelligence? Someone who is going to say We didn’t care enough about this, but now we need to, and will.
From Drexel’s YouTube channel:
But far less recognized is that Drexel made the very bold decision of committing all students to purchase a previously unreleased and untested computer from Apple. This was, of course, the Macintosh (introduced in January 1984), which was unlike any previous computer. Drexel’s commitment to the Mac was also of great benefit to Apple, helping to legitimize this brand-new platform, which helped make the Mac a successful product that continues to thrive in education 40 years later.
This entire initiative, called the Drexel Microcomputer Project, was captured in a 1-hour documentary filmed by David Jones, Dean of the Pennoni Honors College from 2008 to 2014. The film premiered at Drexel in 1985.
I was very fortunate not only to know Dave Jones (who died in 2018) but to have him as a professor for several film criticism courses (one on westerns, and another on the works of Alfred Hitchcock). I was a computer science major, not a film major, but Jones didn’t care. He was also familiar with — dare I say, a fan of — my column in The Triangle, Drexel’s student newspaper. He took me to lunch my senior year and encouraged me to pursue writing as a career. He was a great teacher: thoughtful, kind, insightful, open-minded, and deeply knowledgeable.
I saw a screening of Going National back in 2011, and sat on a panel discussion with Jones to talk about it. It’s a good documentary, and he really captured the feel of Drexel’s campus at the time. It is a very ’80s movie. It was gratifying that I got to tell him, then, that his advice to me back in 1996 had worked out pretty well.
Alissa Falcone, in a good piece looking back at (my alma mater) Drexel University’s groundbreaking deal with Apple 40 years ago to provide deeply discounted Macintoshes to all students, and integrate them throughout the campus and curriculums:
Drexel was prepared to buy IBM computers — and had equipped its computer centers with IBMs for decades — but the cost came to more than $1,000 per unit. IBM’s young competitor Apple, on the other hand, was willing to give discounts, provided the University agreed to secret negotiations and discreet showings of its newest, unreleased personal computer.
Bruce Eisenstein, PhD, Arthur J. Rowland Professor of Electrical and Computer Engineering in the College of Engineering, was the head of the Department of Electrical and Computer Engineering at the time, and had been the founding faculty adviser for the Drexel Computer Society started in 1972. He was Drexel’s choice to meet with an Apple representative to see the future Macintosh, which had never-before-seen properties like a mouse, icons on a screen and different fonts. This new Apple product was more powerful and easier to use than earlier personal computers; novices could supposedly master it in 30 minutes (without the need to memorize and type coded commands). And Apple agreed on the $1,000 price tag for a model that sold to the public for $2,495.
“I went back to the selection committee and I said, ‘Listen, you have to forget the IBM. This new computer from Apple is the one you have to get. They are going to make it available to us for a thousand dollars — that’s all inclusive.’ And the first question was ‘Is it compatible with the IBM computer?’ Well, no. Was there software for it? No. Were there any programs for it, like a word processor? Not yet. So the committee justifiably kept saying, well, what’s the name of this? What’s it like? I couldn’t tell them. I had to say you just gotta trust me on this. So they took a vote and unanimously voted to adopt the unknown computer that turned out to be the Macintosh,” Eisenstein recalled in Building Drexel: The University and Its City, 1891-2016.
Drexel chose the untested Macintosh even knowing that Apple wouldn’t announce it to the public until January 1984 and that the computers wouldn’t be ready until March, almost halfway through that momentous academic year.
I never had the pleasure of meeting Eisenstein, but I’d sure like to thank him for his prescience. By the time I got to Drexel in 1991 the Mac was infused throughout campus.
Ten years ago I played two small roles in the release of the aforelinked Becoming Steve Jobs. First, I got to announce the book here at Daring Fireball, after having been sent an advance copy a few weeks earlier. My praise for the book then was glowing, but in hindsight, I think I undersold just how good — and how essential — it is. At the time of its launch, the book remained in the shadow of Walter Isaacson’s Jobs biography. Ten years later, I refer back to Becoming Steve Jobs regularly; Isaacson’s book almost never.
Second, the SoHo Apple Store in New York hosted a “Meet the Authors” event, and I had the pleasure of playing host for the interview. It’s still available, both as video and audio. And as I wrote at the time, “I’m kind of proud that it got flagged as ‘explicit’ — but that was Bill Gates’s fault.”
During Friday’s episode of Dithering — a free listen — Ben Thompson reminded me that my headline reference last week, alluding to the well-known line from Hamlet, had been used, to great effect, once before. Brent Schlender wrote a crackerjack piece for Fortune in March 1997 under the slightly-different-than-mine headline “Something’s Rotten in Cupertino”.
I should make very clear that I didn’t mean to allude to Schlender’s piece with mine. We both just riffed on the same idiom from Shakespeare. In my case, I’m writing about one initiative that’s gone awry inside a very successful, well-functioning Apple. The timeframe for Schlender’s piece, on the other hand, was the most precarious and dysfunctional period in the history of Apple. Then-CEO Gil Amelio had just announced the acquisition of NeXT in December 1996. By June, most of the board would be replaced, Amelio fired, and Steve Jobs would return as an “advisor”, and, by the end of 1997, as “interim CEO”. Schlender was all over what was really going on:
At Apple’s headquarters in Cupertino, California, a power play is in progress that calls into question who’s really running the company and that may very well put Apple in play once again. So thick is this plot that it reaches into the homes of some of the most powerful CEOs in Silicon Valley. The delicious irony is that what triggered the soap opera is a move Amelio hopes is his masterstroke: Apple’s $400-million acquisition of Next, and the advisory services of Steve Jobs that come bundled with it.
Amelio’s big deal is beginning to look more like a Next takeover of Apple. Never mind that Next Software was a boutique with revenues that would amount to less than a rounding error to Apple. Jobs, the Svengali of Silicon Valley, may have outdone himself this time: Not only did he collect $100 million and 1.5 million shares of Apple stock for his stake in Next, but his fingerprints are all over Amelio’s latest reorganization plan and product strategy — even though Jobs doesn’t have an operational role or even a board seat.
To the Machiavellian eye, it looks as if Jobs, despite the lure of Hollywood — lately he has been overseeing Pixar, maker of Toy Story and other computer-animated films — might be scheming to take over Apple for himself. If anyone doubts he could do it, all you have to do is ask his best friend, Oracle CEO Larry Ellison, the richest man in Silicon Valley. Says he: “Steve’s the only one who can save Apple.”
We all know today this is exactly how it played out. But in March 1997 none of it was obvious at all. It really wasn’t even clear whether Apple would still exist as an independent company by the end of the year. Just extraordinary reporting and storytelling by Schlender.
Go ahead and re-read Schlender’s 1997 Fortune piece, which we should be thankful remains online at all, but which has been mangled, formatting-wise, by whatever series of CMS transitions have kept it online for 28 years. But if you want the best version of this saga, get yourself a copy of Becoming Steve Jobs, the 2015 biography Schlender co-authored with Rick Tetzeli. I just re-read chapter 8 over the weekend — the chapter covering Jobs’s momentous 1997 (in addition to the Apple-NeXT reunification, that was also the year he hammered out an aggressive new deal with Disney for five post-Toy Story feature films) — and it’s just so good. If you don’t already have a copy of Becoming Steve Jobs, get it at Amazon or from Bookshop.org or Apple Books.
A new feature in our membership CMS (Passport — check it out) lets us make individual episodes of Dithering free for everyone to listen to. I can’t think of a better way to first use this new capability than to open up Friday’s episode, recapping my “Something Is Rotten in the State of Cupertino” article, and the resonance with which it hit. Even the cover art — selected weeks ago — captures how I’ve felt this week.
Give it a listen. Subscribe if you enjoy it.
My post Friday commenting (read: wise-cracking) on Mark Gurman’s explosive report on an all-hands Siri team meeting at Apple was begging for a bit of meta commentary on the reporting itself. But I’ve been doing so much of that regarding Gurman lately that I thought it best to hold it for a postscript. Here’s that postscript.
Both of these things are true:
In short, I do actually suspect — but can claim zero sources familiar with the matter to confirm — that Gurman hangs his toilet paper in an improper underhand fashion.
So let’s just examine how extraordinary and singular Gurman’s Friday report was. Nobody else reported on this meeting. Every other article about it — including mine — was commenting on Gurman’s exclusive report about the meeting. I’ve not seen one other report even confirming the meeting took place, let alone describing it in detail, replete with copious quotes from Siri senior director Robby Walker, who, according to Gurman, led the meeting. Not one. I’m not pointing that out to cast suspicion that the meeting did not take place or that Gurman’s report cast it inaccurately or that his direct quotations were not, in fact, direct quotations. I’m pointing out just how singular and extraordinary Mark Gurman is in this sphere. If it wasn’t for Gurman’s report we, outside Apple (and probably outside the Siri team inside Apple) wouldn’t even know the meeting occurred.
How did Gurman not only get the scoop on this meeting, but copious direct quotes from Walker’s remarks to the team? Well, it was “according to people with knowledge of the matter, who asked not to be identified because the gathering was private”. In other words, more than one member of the Siri team, and at least one of which either recorded the meeting surreptitiously and slipped the recording to Gurman, or at least one of whom takes notes at the pace and accuracy of a court stenographer. Either way, these sources — plural — surely knew how the meeting would make Apple look if it were to leak.
I’ve long made my opinions about Bloomberg’s institutional journalistic credibility well known. But I don’t think they’re bereft of credibility — it’s the fact that they are deservedly well-regarded that makes their refusal to ever admit their own glaring mistakes so notable. When a Gurman reports says “people” that means “more than one” and, I believe, he must be able to confirm to his editors that he got this information from more than one source. If he’s reporting direct quotes, I think that means he’s heard a recording. That’s extraordinary.
But I’d feel a lot better about our collective conventional wisdom regarding the nature of this particular all-hands Siri meeting if it had leaked to, and been reported on by, more than one reporter at more than one publication. ★
My thanks to WorkOS for sponsoring this week at DF. Modern authentication should be seamless and secure. WorkOS makes it easy to integrate features like MFA, SSO, and RBAC.
Whether you’re replacing passwords, stopping fraud, or adding enterprise auth, WorkOS can help you build frictionless auth that scales. Future-proof your authentication stack with the identity layer trusted by OpenAI, Cursor, Perplexity, and Vercel. Upgrade your auth today.
From Apple’s support documentation:
You can generate a report of requests your iPhone has sent to Private Cloud Compute.
Go to Settings, then tap Privacy & Security.
Tap Apple Intelligence Report, then choose a report duration for the last 15 minutes (default) or last 7 days. Choose off to disable the report.
Note: The report may be empty if there haven’t been any Private Cloud Compute requests since you changed the duration.
Tap Export Activity, choose a place to store the file, then tap Export.
The report is saved as a file named Apple_Intelligence_Report.json.
Open the file with a text reader.
These are the iOS instructions, but they’re exactly the same on MacOS 15 Sequoia. My first generated report was empty for the last 7 days, and it was empty again even after running the Writing Tools Proofread function on the text of my 4,000-word “Something Is Rotten in the State of Cupertino” article from this week. But when I ran the Writing Tools Summarize feature on the same text, I wound up with a long entry that was sent to Private Cloud Compute. So, at the moment, Summarize seems like a good way to invoke Private Cloud Compute, even from a relatively powerful Mac.
Here’s the summary Apple Intelligence generated. I have to say: it’s pretty good. It’s completely petty but also completely me to notice and object to the way it uses two spaces after periods — and worse, only some of the time. Also, the sentence “This raises concerns about the company’s ability to maintain its position as a leader in AI innovation” is, let’s say, off the mark.
Update: Howard Oakley wrote a post with a brief overview of the structure and contents of these reports back on October 29.
Jess Weatherbed, reporting for The Verge:
iPhone and Android users will be able to exchange end-to-end encrypted (E2EE) RCS messages in the near future thanks to newly updated RCS specifications. The GSM Association announced that the latest RCS standard includes E2EE based on the Messaging Layer Security (MLS) protocol, enabling interoperable encryption between different platform providers for the first time. [...]
“End-to-end encryption is a powerful privacy and security technology that iMessage has supported since the beginning, and now we are pleased to have helped lead a cross industry effort to bring end-to-end encryption to the RCS Universal Profile published by the GSMA,” said Apple spokesperson Shane Bauer. “We will add support for end-to-end encrypted RCS messages to iOS, iPadOS, macOS, and watchOS in future software updates.” [...]
“We’ve always been committed to providing a secure messaging experience, and Google Messages users have had end-to-end encrypted (E2EE) RCS messaging for years,” Google spokesperson Ed Fernandez told The Verge. “We’re excited to have this updated specification from GSMA and work as quickly as possible with the mobile ecosystem to implement and extend this important user protection to cross-platform RCS messaging.”
This is nothing but good news. But it’s wrong to frame this along Google’s lines, that they’ve been there waiting for Apple to support E2EE for RCS. They’ve been waiting for Apple to support RCS at all, yes, and Google has also implemented their own proprietary E2EE layer for RCS. But until now, there was no E2EE specification in the open RCS spec. Now there is. That’s why it’s not just Android ↔︎ iOS RCS messaging that wasn’t able to use E2EE, but even Android ↔︎ Android, unless both devices were using Google’s own Messages app.
I have also noticed recently that Google Messages and Apple Messages now do a pretty good job of supporting each other’s tapbacks. And that hasn’t done anything to really change the green/blue messaging dynamic. Both things are true: RCS makes cross-platform messaging way better and iMessage remains vastly superior to RCS.
What I’m most interested about with Apple’s implementation of RCS encryption is how they’ll indicate it visually in chats. It’s not going to be with blue bubbles. Blue means “iMessage”, not “encrypted” — it just happens to be that iMessage started as a protocol based on end-to-end encryption. There’s no such thing as a non-encrypted iMessage — it’s part of the protocol, and always has been. But what happens when new/updated Android phones support the new RCS encryption spec, and older devices don’t? A lock icon for the encrypted chats? If it were up to me, iOS would drop support for non-encrypted RCS — iOS should use RCS with E2EE for every device that supports it, and fall back to dumb old no-encryption-at-all SMS for all devices that do not.
In an item earlier this week observing that Swift Assist, the most ambitious Xcode-related Apple Intelligence feature shown at WWDC last year, not only hasn’t yet shipped but still is not in beta, I wondered whether Apple actually demoed it live last year. John Voorhees, writing for MacStories from WWDC last June, reports that they did:
Earlier today, I got the very first live demo of Swift Assist, one of the many developer tools introduced today by Apple. I also saw code completion in action. It was an impressive demo, and although the tools seem like magic and will undoubtedly be valuable to developers, they do have their limitations, which are worth exploring. [...]
The code completion demo also included a live demo of Swift Assist. Unlike code completion, Swift Assist requires an Internet connection because requests are sent to the cloud. As a result, it takes several seconds for Swift Assist to return results. The delay was noticeable compared to the speed of code completion, but it wasn’t a painfully long wait either.
I heard this week from a third-party developer who was invited to Apple for a one-day hands-on session with Swift Assist late last year. Swift Assist was definitely working, but seemingly not working too well. From that source: “The UI is very much complete (just like Siri), but the results the LLM produces were not very good. It could make very basic demo apps with a prompt like ‘make an app that takes the NASA satellite JSON and shows the current satellites traveling overhead right now’, but not too much more than that. It fell apart on more complex tasks.”
I remember the remote-inference-only aspect of the Swift Assist presentation from my Xcode briefing at WWDC: that because of its complexity, Swift Assist would not execute locally and would only run via Private Cloud Compute. My own notes on this from WWDC were mostly related to the privacy and security implications. That developers should feel safe using Swift Assist even with confidential code and projects because Private Cloud Compute would be guaranteed private. I also remember thinking, at the time, that I should be more skeptical about Apple’s claims about Apple Intelligence features that would execute locally, on-device, rather than the ones that would execute remotely, via Private Cloud Compute, because the way almost all “AI” features from other companies over the previous two years worked was entirely in the cloud. Apple’s statements that Apple Intelligence will perform much inference locally, on-device, seemed like the stretch goal.
But now in March 2025 I’m beginning to think it’s the other way around. What features and aspects of Apple Intelligence run in Private Cloud Compute, today, in March 2025? Do any? I’ve been poking around for a few days and I don’t have any answers. Is Private Cloud Compute running in production yet? How would we know? If you know, let me know.
Update: “How to Generate a Report of Apple Intelligence Requests Sent to Private Cloud Compute”.
Katie Notopoulos, writing at Business Insider (Apple News+ link):
It’s possible that this strident defense is backfiring — creating a “Streisand Effect” that’s publicizing the book even more.
Notopoulos isn’t one to pull punches or hedge, normally, but there’s no question that Meta has Streisand Effected the hell out of Wynn-Williams’s Careless People. I’m on my way to buy a copy this afternoon and I hadn’t even heard of its existence a day ago.
Sarah Wynn-Williams worked at Facebook from 2011 to 2017 — and her book, “Careless People,” details what she said were a bunch of bad things the company did. It also contains allegations that Joel Kaplan — who is now Meta’s chief global affairs officer — sexually harassed her. (Meta said this week that Kaplan had been cleared of the harassment allegations in 2017 after it investigated Wynn-Williams’ complaint.)
The book was released with hardly any pre-publishing fanfare: It was announced by its publisher, Flatiron Books, an imprint of Macmillan, only a few days before it came out Tuesday. That’s an unusually short timeline.
Meta’s reaction has been emphatic and multi-pronged. The company created a page on its website that detailed a series of the claims in the book. It said some of these — like claims about its ambitions in China or its alleged failure to act in Myanmar — have already been reported in the news media as far back as 2017 and have been acknowledged by the company publicly. The company lists seven “new” claims from the book and then offers links to past coverage and company responses — calling the claims “old news.”
Meta’s strongest move was on Wednesday when it won a ruling in arbitration that said Wynn-Williams could no longer promote the book because of a non-disparagement clause in a contract she signed as an employee. The ruling was granted on an emergency basis and is temporary pending the completion of the full arbitration process, The Washington Post reported.
“Read the book that Meta not only doesn’t want you to read, but wants to prevent its author from even talking about” is a hell of a marketing angle. What an own goal.
Rick Porter, The Hollywood Reporter:
After more than a year of speculation, Apple TV+ has ordered a fourth season of its Emmy-winning comedy Ted Lasso. The pickup comes after co-creator and star Jason Sudeikis closed a deal to reprise his role as the title character, the manager of the fictional AFC Richmond soccer club in London.
I dig Ted Lasso but this is like 0.1 percent as interesting as When are we going to get Severance season 3?! Gimme gimme gimme. And we haven’t even gotten to next week’s season 2 finale yet.
CNBC story from 2016:
At a recent public appearance at the Utah Tech Tour, in a conversation moderated by Utah’s Senator Orrin Hatch, Apple CEO Tim Cook pointed out that Microsoft had tablets on the market decades before Apple.
Cook emphasized his company’s timing coming to market with new products to underscore the idea that it’s nearly impossible for a company to be the best, the first and to make the most of a given product. [...]
“It doesn’t bother us that we are second, third, fourth or fifth if we still have the best. We don’t feel embarrassed because it took us longer to get it right,” says Cook.
“For Apple, being the best is the most important and trumps the other two by far.”
This has been one of Apple’s guiding mantras for decades, and it has served the company very well. But it stops holding water when they promise to be first, but then aren’t first and aren’t the best.
If you only ever promise A, B, and C — and never mention X, Y, or Z — even when competitors ship their versions of X, Y, and Z first, your silence speaks for itself. Either you don’t think X, Y, and Z are important, or, you think it’s worth taking more time to get them right. But if you promise A, B, C, X, Y, and Z, and then only ship A, B, and C, you just look lost when competitors ship X, Y, and Z.
Jay Peters, last month for The Verge:
Google’s next I/O developer conference will take place on May 20th and May 21st, the company announced today. The event will be “open to everyone online” and will include “livestreamed keynotes and sessions,” according to an FAQ. Like past years, there will also be an in-person component at the Shoreline Amphitheatre in Mountain View, California.
That’s still two months off but I got interested in the dates for I/O this week. After we recorded Friday’s Dithering (on which we talked about my “Something Is Rotten in the State of Cupertino” piece and, more so, the reaction to it, the resonance it seemed to strike), Ben Thompson and I were spitballing, and it popped into my head that Apple’s “more personalized Siri” delay is a marketing gift to Google.
What I would do if I worked at Google is prepare a live demo of Google Gemini on a Pixel phone doing exactly what Apple showed in last year’s announcement at WWDC, and then again in the Bella Ramsey TV commercial that Apple pulled from YouTube. Something like this:
Presenter: This is a live demo, on my Pixel 9. I need to pick my mom up at the airport and she sent me an email with her flight information. [Invokes Gemini on phone in hand...] Gemini, when is my mom’s flight landing?
Gemini: Your mom’s flight is on time, and arriving at SFO at 11:30.
Presenter: I don’t always remember to add things to my calendar, and so I love that Gemini can help me keep track of plans that I’ve made in casual conversation, like this lunch reservation my mom mentioned in a text. [Invokes Gemini...] What’s our lunch plan?
Gemini: You’re having lunch at Waterbar at 12:30.
Presenter: How long will it take us to get there from the airport?
Gemini presents a Google Maps driving directions popup window showing it will take 21 minutes.
Then do another live demo with the “What’s the name of the guy I had a meeting with a couple of months ago at Cafe Grenel?” example from Apple’s now-pulled Apple Intelligence commercial. The exact same demos, but real: live and on stage. These would be great demos even if Apple had never promised to deliver them. But given that Apple did promise them for this year, and has now delayed them until “the coming year”, they’re devastating if Google can show them actually working on Apple’s own original timeline.
Stick to what Gemini can actually do, on actual Pixel phones running the new beta software — which by some accounts already includes the delayed personalized features of Siri. The point should be showing Google AI technology, accessing personal data from Google cloud services, using Google devices, delivering on the promises made by Apple a year prior. But there’s no need for Google’s presenters to mention Apple or Siri, or even mention unnamed “competitors”. Everyone watching the I/O keynote will recognize those demo prompts and draw the competitive conclusions for themselves. Nothing comes across as confident like not even acknowledging, let alone naming, your competition. And nothing serves as proof like a live working demo.
Mark Gurman, reporting for Bloomberg today:
Apple Inc.’s top executive overseeing its Siri virtual assistant told staff that delays to key features have been ugly and embarrassing, and a decision to publicly promote the technology before it was ready made matters worse.
Robby Walker, who serves as a senior director at Apple, delivered the stark comments during an all-hands meeting for the Siri division, saying that the team was facing a bad period. Walker also said that it’s unclear when the enhancements will actually launch, according to people with knowledge of the matter, who asked not to be identified because the gathering was private.
Robby Walker is not Apple’s “top executive overseeing its Siri virtual assistant”. Take your pick of whether that’s SVP John Giannandrea or CEO Tim Cook, but Walker reports to Giannandrea. Gurman, of course, knows this better than I do; I suspect he knows Apple’s entire org chart. But it makes Bloomberg’s headline — “Apple’s Siri Chief Calls AI Delays Ugly and Embarrassing, Promises Fixes” — misleading. Also somewhat misleading in that headline is that these comments from Walker were clearly not meant to leak. This is not a public apology, like the one Tim Cook wrote and signed in 2012 in the aftermath of the Apple Maps launch with iOS 6.
Still, he praised the team for developing “incredibly impressive” features and vowed to deliver an industry-leading virtual assistant to consumers.
Those two words are the only direct quote in the first nine paragraphs of Bloomberg’s report, which is kind of crazy, because the second half of Gurman’s story is full of quotes. I suspect his editors did Gurman a disservice on this one. The quotes are juicy AF, but don’t really start until the 10th paragraph. Like, for example, which Siri features does Walker think are “incredibly impressive”? No snark, I’d love to know. Is it Siri’s sports knowledge? The new product knowledge feature, that gives incomplete and/or incorrect instructions for how to toggle preferences in Settings? OK I guess that’s some snark, but I sincerely and honestly would love to know which Siri features the senior director in charge of Siri considers “incredibly impressive”.
Walker told staff in the meeting that the delays were especially “ugly” because Apple had already showed off the features publicly. “This was not one of these situations where we get to show people our plan after it’s done,” he said. “We showed people before.”
“To make matters worse,” Walker said, Apple’s marketing communications department wanted to promote the enhancements.
Again, this meeting clearly was not intended to leak. (It’s perhaps another knock against the Siri team, in addition to the quality of their output, that they leak internal meetings.) But that comes across as Walker blaming marketing.
Walker also raised doubts about even meeting the current release expectations. Though Apple is aiming for iOS 19, it “doesn’t mean that we’re shipping then,” Walker said. The company has several more priorities in development, and trade-offs will need to be made, he said.
“We have other commitments across Apple to other projects,” Walker said, citing new software and hardware initiatives. “We want to keep our commitments to those, and we understand those are now potentially more timeline-urgent than the features that have been deferred.” He said decisions on timing will be made on a “case-by-case basis” as work progresses on products planned for next year.
“Customers are not expecting only these new features but they also want a more fully rounded-out Siri,” he said. “We’re going to ship these features and more as soon as they are ready.”
The customers still haven’t gotten their appetizers, but it’s time to start the entrees, so the kitchen staff is working on those now. But don’t worry, they hope to get the appetizers out alongside dessert. But it’s OK because dessert might be late too.
As of Friday, Apple doesn’t plan to immediately fire any top executives over the AI crisis, according to people with knowledge of the matter.
I’d be very curious to know just how many people could possibly be familiar with this particular matter. Because this particular matter comes down to what Tim Cook is thinking. I’m thinking it’s about six people he might discuss this with. Wait, no, I just thought of a seventh. I’m going to say seven, tops. But maybe I’m wrong, and Tim Cook is the chatty sort, who openly talks with a large number of senior managers about whom he might fire.
Walker said the decision to delay the features was made because of quality issues and that the company has found the technology only works properly up to two-thirds to 80% of the time. He said the group “can make more progress to get those percentages up, so that users get something they can really count on.”
It’s unclear exactly which features these are in reference to, but presumably they’re not the “incredibly impressive” ones. Because something that “only works properly up to two-thirds” of the time only seems regular impressive to me, not incredibly impressive. Maybe it’s the ones that work properly four out of five times that are incredible?
Walker compared the endeavor to an attempt to swim to Hawaii. “We swam hundreds of miles — we set a Guinness Book for World Records for swimming distance — but we still didn’t swim to Hawaii,” he said. “And we were being jumped on, not for the amazing swimming that we did, but the fact that we didn’t get to the destination.”
I’d say it’s a little more like selling customers tickets for a cruise that includes a stop in Hawaii, then never actually getting to Hawaii, and hoping they didn’t notice when the ship returns to port to disembark.
He showed examples during the meeting of the technology working: It was able to locate his driver’s license number on command and find specific photos of a child. He also demonstrated how the technology could precisely manipulate apps via voice control. It embedded content in an email, added recipients and made other changes.
That’s the biggest actual news in the report, and it’s in paragraph 23.
Walker said that some staffers may feel “relieved” over the delays. “If you were using these features in the build, you were probably wondering: Are these ready? How do I feel about shipping these to our customers? Is this the right choice?”
He added that some employees “might be feeling embarrassed.”
Again, I’ll reiterate that this was a private meeting, not meant to leak. Maybe it’s an inaccurate summary. I hope it is. But as reported by Gurman, this meeting reeks of you-all-deserve-participation-trophies-to-reward-your-hard-work-and-it’s-OK-to-feel-embarrassed vibes. What’s needed, quite obviously, is some “What is it supposed to do? / So why the fuck doesn’t it do that?” vibes.
As a postscript to that last item, it occurred to me that because we’re close friends, I have a lot of photos of Paul Kafasis in my library. Here’s one from a year ago you can use as a reference. I wondered how Genmoji would do with “An owl who looks like Paul Kafasis, wearing a Celtics jersey, holding a basketball.” Here’s the best one, from my Mac running MacOS 15.3.1 Sequoia.
It does resemble Paul. Not sure about how owl-like it is, the “holding” part, the slopping of “Celtics”, or the quality of orthodontia over in Genmojiworld.
Paul Kafasis, on a seemingly local-to-Boston Genmoji billboard from Apple:
Eventually, though, the penny dropped. After my umpteenth time passing the billboard, while trying to distract myself from the single-digit temperatures and the brutal wind chill, I realized what I’d been missing. That’s not just a basketball-playing
foulfowl.You see, the Last Tenement is located just a few hundred feet from TD Garden, home arena for the reigning NBA champion Boston Celtics. The Celtics have had many great players over the years, but only one had the last name “Bird”. Yes, this billboard is surely an allusion to the Hick from French Lick, Larry Bird.
It was obvious once I finally saw it, and I was so amused that I went by with my phone to snap the above pic. At that point, I noticed that the cartoon’s basketball jersey even seems to feature Larry Legend’s #33. It’s cut off, perhaps in the hopes of avoiding a lawsuit, but those numbers really can’t be anything else.
I can’t believe Kafasis didn’t recognize the caricature immediately, but he’s a bit younger than me and didn’t grow up like I did, familiarizing myself with the early 1980s “Familiar Birds of North America” illustrated guide from ... checks notes ... Playboy.
Zack Whittaker, reporting for TechCrunch:
A group of bipartisan U.S. lawmakers are urging the head of the U.K.’s surveillance court to hold an open hearing into Apple’s anticipated challenge of an alleged secret U.K. government legal demand.
U.S. Senator Ron Wyden, along with four other federal lawmakers, said in a letter this week to the president of the U.K.’s Investigatory Powers Tribunal (IPT) that it is “in the public interest” that any hearings about the alleged order are not held in secret.
From the letter, signed by Senators Ron Wyden (D-OR) and Alex Padilla (D-CA), and Congresspeople Zoe Lofgren (D-CA), Andy Biggs (R-AZ), and Warren Davidson (R-OH):
We write to request the Investigatory Powers Tribunal (IPT) remove the cloak of secrecy related to notices given to American technology companies by the United Kingdom, which infringes on free speech and privacy, undermines important United States Congress and U.K. parliamentary oversight, harms national security, and ultimately, undermines the special relationship between the United States and the United Kingdom. [...]
Given the significant technical complexity of this issue, as well as the important national security harms that will result from weakening cybersecurity defenses, it is imperative that the U.K.’s technical demands of Apple — and of any other U.S. companies — be subjected to robust, public analysis and debate by cybersecurity experts. Secret court hearings featuring intelligence agencies and a handful of individuals approved by them do not enable robust challenges on highly technical matters. Moreover, given the potential impact on U.S. national security, it is vital that American cybersecurity experts be permitted to analyze and comment on the security of what is proposed.
Michael Tsai:
Swift Assist was supposed to arrive in 2024, but it never even appeared in a beta. Apple hasn’t announced that it’s postponed or cancelled. It’s not even mentioned in the release notes.
Apple announced two AI-powered features for Xcode last year: predictive code completion and Swift Assist. Predictive code completion is basically getting suggestions for what you might intend to write next. Swift Assist is far more ambitious, runs only in the cloud (via Private Cloud Compute), and promises to write entire components for you. So much like the rest of Apple Intelligence, it’s the simpler thing that has shipped, and the more radical ambitious thing that has not.
I looked into Swift Assist this week while researching, and checking my notes from WWDC, for my long piece yesterday. I had an entire briefing with the Xcode team at WWDC last year, and alas, I was late for it (previous briefing, with a different team, had started late), so I missed some stuff and my notes are unusually terse. What I wanted to check my notes for is whether or not I was shown a live demo of Swift Assist last June. But my notes don’t say, and my own memory is completely unsure whether the demos I saw performed were only for predictive code completion, or included Swift Assist.
If anyone else who was in those WWDC briefings remembers whether Swift Assist was actually demoed, please let me know. I’m genuinely curious if Swift Assist was another thing — like all of “more personalized Siri” — that wasn’t even in demonstratable shape at WWDC.
Update, 15 March: “Apple Did Demo Swift Assist at WWDC Last Year, and Has Shown It, Under NDA, Since Then”.
In the two decades I’ve been in this racket, I’ve never been angrier at myself for missing a story than I am about Apple’s announcement on Friday that the “more personalized Siri” features of Apple Intelligence, scheduled to appear between now and WWDC, would be delayed until “the coming year”.
I should have my head examined.
This announcement dropped as a surprise, and certainly took me by surprise to some extent, but it was all there from the start. I should have been pointing out red flags starting back at WWDC last year, and I am embarrassed and sorry that I didn’t see what should have been very clear to me from the start.
How I missed this is twofold. First, I’d been lulled into complacency by Apple’s track record of consistently shipping pre-announced products and features. Their record in that regard wasn’t perfect, but the exceptions tended to be around the edges. (Nobody was particularly clamoring for Apple to make a multi-device inductive charging mat, so it never generated too much controversy when AirPower turned out to be a complete bust.) Second, I was foolishly distracted by the “Apple Intelligence” brand umbrella. It’s a fine idea for Apple to brand its AI features under an umbrella term like that, similar to how a bunch of disparate features that allow different Apple devices to interoperate are under the “Continuity” umbrella. But there’s no such thing, technically speaking, as “Continuity”. It’s not like there’s an Xcode project inside Apple named Continuity.xcodeproj, and all the code that supports everything from AirDrop to Sidecar to iPhone Mirroring to clipboard sharing is all implemented in the same framework of code. It’s a marketing term, but a useful one — it helps Apple explain the features, and helps users understand them.
The same goes for “Apple Intelligence”. It doesn’t exist as a single thing or project. It’s a marketing term for a collection of features, apps, and services. Putting it all under a single obvious, easily remembered — and easily promoted — name makes it easier for users to understand that Apple is launching a new initiative. It also makes it easier for Apple to just say “These are the devices that qualify for all of these features, and other devices — older ones, less expensive ones — get none of them.”
Let’s say Apple were to quietly abandon the dumb Image Playground app next year. It just disappears from iOS 19 and MacOS 16. That would just be Apple eliminating a silly app that almost no one uses or should use. That wouldn’t be a setback or rollback of “Apple Intelligence”. I would actually argue that axing Image Playground would improve Apple Intelligence; its mere existence greatly lowers the expectations for how good the whole thing is.1
What I mean by that is that it was clear to me from the WWDC keynote onward that some of the features and aspects of Apple Intelligence were more ambitious than others. Some were downright trivial; others were proposing to redefine how we will do our jobs and interact with our most-used devices. That was clear. But yet somehow I didn’t focus on it. Apple itself strongly hinted that the various features in Apple Intelligence wouldn’t all ship at the same time. What they didn’t spell out, but anyone could intuit, was that the more trivial features would ship first, and the more ambitious features later. That’s where the red flags should have been obvious to me.
In broad strokes, there are four stages of “doneness” or “realness” to features announced by any company:
Features that the company’s own product representatives will demo, themselves, in front of the media. Smaller, more personal demonstrations are more credible than on-stage demos. But the stakes for demo fail are higher in an auditorium full of observers.
Features that the company will allow members of the media (or other invited outside observers and experts) to try themselves, for a limited time, under the company’s supervision and guidance. Vision Pro demos were like this at WWDC 2023. A bunch of us got to use pre-release hardware and in-progress software for 30 minutes. It wasn’t like free range “Do whatever you want” — it was a guided tour. But we were the ones actually using the product. Apple allowed hands-on demos for a handful of media (not me) at Macworld Expo back in 2007 with prototype original iPhones — some of the “apps” were just screenshots, but most of the iPhone actually worked.
Features that are released as beta software for developers, enthusiasts, and the media to use on their own devices, without limitation or supervision.
Features that actually ship to regular users, and hardware that regular users can just go out and buy.
As of today — March 2025 — every feature in Apple Intelligence that has actually shipped was at level 1 back at WWDC. After the keynote, dozens of us in the press were invited to a series of small-group briefings where we got to watch Apple reps demo features like Writing Tools, Photos Clean Up, Genmoji, and more. We got to see predictive code completion in Xcode. What has shipped, as of today, they were able to show, in some functional state, in June.
For example, there was a demo involving a draft email message on an iPad, and the Apple rep used Writing Tools to make it “more friendly”. I was in a group of just four or five other members of the media, watching this. As usual, we were encouraged to interrupt with questions. Knowing that LLMs are non-deterministic, I asked whether, as the Apple rep was performing this same demo for each successive group of media members, the “more friendly” result was exactly the same each time. He laughed and said no — that while the results are very similar each time, and he hopes they continue to be (hence the laughing), there were subtle differences sometimes between different runs of the same demo. As I recall, he even used Undo to go back to the original message text, invoked Writing Tools to make it “more friendly” again, and we could see that a few of the word choices were slightly different. That answered both my explicit question and my implicit one: Writing Tools generates non-deterministic results, and, more importantly, what we were watching really was a live demo.
We didn’t get to try any of the Apple Intelligence features ourselves. There was no Apple Intelligence “hands on”. But we did see a bunch of features demoed, live, by Apple folks. In my above hierarchy of realness, they were all at level 1.
But we didn’t see all aspects of Apple Intelligence demoed. None of the “more personalized Siri” features, the ones that Apple, in its own statement announcing their postponement, described as having “more awareness of your personal context, as well as the ability to take action for you within and across your apps”. Those features encompass three main things:
“In-app actions” — Giving Siri the ability, through the App Intents framework, to do things in and across apps that you can do, the old fashioned way (yourself) in and across apps. Again, here’s Apple’s own example usage:
You can make a request like “Send the email I drafted to April and Lilly” and Siri knows which email you’re referencing and which app it’s in. And Siri can take actions across apps, so after you ask Siri to enhance a photo for you by saying “Make this photo pop,” you can ask Siri to drop it in a specific note in the Notes app — without lifting a finger.
There were no demonstrations of any of that. Those features were all at level 0 on my hierarchy. That level is called vaporware. They were features Apple said existed, which they claimed would be shipping in the next year, and which they portrayed, to great effect, in the signature “Siri, when is my mom’s flight landing?” segment of the WWDC keynote itself, starting around the 1h:22m mark. Apple was either unwilling or unable to demonstrate those features in action back in June, even with Apple product marketing reps performing the demos from a prepared script using prepared devices.
This shouldn’t have just raised a concern in my head. It should have set off blinding red flashing lights and deafening klaxon alarms.
Even the very engineers working on a project never know exactly how long something is going to take to complete. An outsider observing a scripted demo of incomplete software knows far less (than the engineers) just how much more work it needs. But you can make a rough judgment. And that’s where my aforementioned hierarchy of realness comes into play. Even outsiders can judge how close a public beta (stage 3) feels to readiness. A feature or product that Apple will allow the press to play with, hands-on (stage 2) is further along than a feature or product that Apple is only willing to demonstrate themselves (stage 1).
But a feature or product that Apple is unwilling to demonstrate, at all, is unknowable. Is it mostly working, and close to, but not quite, demonstratable? Is it only kinda sorta working — partially functional, but far from being complete? Fully functional but prone to crashing — or in the case of AI, prone to hallucinations and falsehoods? Or is it complete fiction, just an idea at this point?
What Apple showed regarding the upcoming “personalized Siri” at WWDC was not a demo. It was a concept video. Concept videos are bullshit, and a sign of a company in disarray, if not crisis. The Apple that commissioned the futuristic “Knowledge Navigator” concept video in 1987 was the Apple that was on a course to near-bankruptcy a decade later. Modern Apple — the post-NeXT-reunification Apple of the last quarter century — does not publish concept videos. They only demonstrate actual working products and features.
Until WWDC last year, that is.
My deeply misguided mental framework for “Apple Intelligence” last year at WWDC was something like this: Some of these features are further along than others, and Apple is showing us those features in action first, and they will surely be the features that ship first over the course of the next year. The other features must be coming to demonstratable status soon. But the mental framework I should have used was more like this: Some of these features are merely table stakes for generative AI in 2024, but others are ambitious, groundbreaking, and, given their access to personal data, potentially dangerous. Apple is only showing us the table-stakes features, and isn’t demonstrating any of the ambitious, groundbreaking, risky features.
It gets worse. Come September, Apple held its annual big event at Apple Park to unveil the iPhone 16 lineup. Apple Intelligence features were highlighted in the announcement. Members of the media from around the world were gathered. That was a new opportunity, three months after WWDC, for Apple to demonstrate — or even better, offer hands-on access to the press to try themselves — the new personalized Siri features. They did not. No demos, at all. But they did promote them, once again, in the event keynote.2
But yet while Apple still wouldn’t demonstrate these features in person, they did commission and broadcast a TV commercial showing these purported features in action, presenting them as a reason to purchase a new iPhone — a commercial they pulled, without comment, from YouTube this week.
Last week’s announcement — “It’s going to take us longer than we thought to deliver on these features and we anticipate rolling them out in the coming year” — was, if you think about it, another opportunity to demonstrate the current state of these features. Rather than simply issue a statement to the media, they could have invited select members of the press to Apple Park, or Apple’s offices in New York, or even just remotely over a WebEx conference call, and demonstrated the current state of these features live, on an actual device. That didn’t happen. If these features exist in any sort of working state at all, no one outside Apple has vouched for their existence, let alone for their quality.
Why did Apple show these personalized Siri features at WWDC last year, and promise their arrival during the first year of Apple Intelligence? Why, for that matter, do they now claim to “anticipate rolling them out in the coming year” if they still currently do not exist in demonstratable form? (If they do exist today in demonstratable form, they should, you know, demonstrate them.)
I’m not trying to be obtuse here. It’s obvious why some executives at Apple might have hoped they could promote features like these at WWDC last year. Generative AI is the biggest thing to happen in the computer industry since previous breakthroughs this century like mobile (starting with the iPhone, followed by Android), social media (Meta), and cloud computing (Microsoft, Google, and Amazon). Nobody knows where it’s going but wherever it’s heading, it’s going to be big, important, and perhaps profitable. Wall Street certainly noticed. And prior to WWDC last year, Apple wasn’t in the game. They needed to pitch their AI story. And a story that involved nothing but table-stakes AI features isn’t nearly as compelling a story as one that involves innovative, breakthrough, ambitious personal features.
But while there’s an obvious appeal to Apple pitching the most compelling, most ambitious AI story possible, the only thing that was essential was telling a story that was true. If the truth was that Apple only had features ready to ship in the coming year that were table stakes compared to the rest of the industry, that’s the story they needed to tell. Put as good a spin on it as possible, but them’s the breaks when you’re late to the game.
The fiasco here is not that Apple is late on AI. It’s also not that they had to announce an embarrassing delay on promised features last week. Those are problems, not fiascos, and problems happen. They’re inevitable. Leaders prove their mettle and create their legacies not by how they deal with successes but by how they deal with — how they acknowledge, understand, adapt, and solve — problems. The fiasco is that Apple pitched a story that wasn’t true, one that some people within the company surely understood wasn’t true, and they set a course based on that.
The Apple of the Jobs exile years — the Sculley / Spindler / Amelio Apple of 1987–1997 — promoted all sorts of amazing concepts that were no more real than the dinosaurs of Jurassic Park, and promised all sorts of hardware and (especially) software that never saw the light of day. Promoting what you hope to be able to someday ship is way easier and more exciting than promoting what you know is actually ready to ship. However close to financial bankruptcy Apple was when Steve Jobs returned as CEO after the NeXT reunification, the company was already completely bankrupt of credibility. Apple today is the most profitable and financially successful company in the history of the world. Everyone notices such success, and the corresponding accumulation of great wealth. Less noticed, but to my mind the more impressive achievement, is that over the last three decades, the company also accumulated an abundant reserve of credibility. When Apple showed a feature, you could bank on that feature being real. When they said something was set to ship in the coming year, it would ship in the coming year. In the worst case, maybe that “year” would have to be stretched to 13 or 14 months. You can stretch the truth and maintain credibility, but you can’t maintain credibility with bullshit. And the “more personalized Siri” features, it turns out, were bullshit.
Keynote by keynote, product by product, feature by feature, year after year after year, Apple went from a company that you couldn’t believe would even remain solvent, to, by far, the most credible company in tech. Apple remains at no risk of financial bankruptcy (and in fact remains the most profitable company in the world). But their credibility is now damaged. Careers will end before Apple might ever return to the level of “if they say it, you can believe it” credibility the company had earned at the start of June 2024.
Damaged is arguably too passive. It was squandered. This didn’t happen to Apple. Decision makers within the company did it.
Who decided these features should go in the WWDC keynote, with a promise they’d arrive in the coming year, when, at the time, they were in such an unfinished state they could not be demoed to the media even in a controlled environment? Three months later, who decided Apple should double down and advertise these features in a TV commercial, and promote them as a selling point of the iPhone 16 lineup — not just any products, but the very crown jewels of the company and the envy of the entire industry — when those features still remained in such an unfinished or perhaps even downright non-functional state that they still could not be demoed to the press? Not just couldn’t be shipped as beta software. Not just couldn’t be used by members of the press in a hands-on experience, but could not even be shown to work by Apple employees on Apple-controlled devices in an Apple-controlled environment? But yet they advertised them in a commercial for the iPhone 16, when it turns out they won’t ship, in the best case scenario, until months after the iPhone 17 lineup is unveiled?
When that whole campaign of commercials appeared, I — along with many other observers — was distracted by the fact that none of the features in Apple Intelligence had yet shipped. It’s highly unusual, and arguably ill-considered, for Apple to advertise any features that haven’t yet shipped. But one of those commercials was not at all like the others. The other commercials featured Apple Intelligence features that were close to shipping. We know today they were close to shipping because they were either in the iOS 18.1 betas already, in September, or would soon appear in developer betas for iOS 18.2 and 18.3. Right now, today, they’ve all actually shipped and are in the hands of iPhone 16 users. But the “Siri, what’s the name of the guy I had a meeting with a couple of months ago at Cafe Grenel?” commercial was entirely based on a feature Apple still has never even demonstrated.
Who said “Sure, let’s promise this” and then “Sure, let’s advertise it”? And who said “Are you crazy, this isn’t ready, this doesn’t work, we can’t promote this now?” And most important, who made the call which side to listen to? Presumably, that person was Tim Cook.
Even with everything Apple overpromised (if not outright lied about) at the WWDC keynote, the initial takeaway from WWDC from the news media was wrongly focused on their partnership with OpenAI. The conventional wisdom coming out of the keynote was that Apple had just announced something called “Apple Intelligence” but it was powered by ChatGPT, when in fact, the story Apple told was that they — Apple — had built an entire system called Apple Intelligence, entirely powered by Apple’s own AI technology, and that it spanned from on-device execution all the way to a new Private Cloud Compute infrastructure they not only owned but are powering with their own custom-designed server hardware based on Apple Silicon chips. And that on top of all that, as a proverbial cherry on top, Apple also was adding an optional integration layer with ChatGPT.
So, yes, given that the news media gave credit for Apple’s own actual announced achievements to OpenAI, Apple surely would have been given even less credit had they not announced the “more personalized Siri” features. It’s easy to imagine someone in the executive ranks arguing “We need to show something that only Apple can do.” But it turns out they announced something Apple couldn’t do. And now they look so out of their depth, so in over their heads, that not only are they years behind the state-of-the-art in AI, but they don’t even know what they can ship or when. Their headline features from nine months ago not only haven’t shipped but still haven’t even been demonstrated, which I, for one, now presume means they can’t be demonstrated because they don’t work.
In May 2011, Fortune published an extraordinary look inside Apple by Adam Lashinsky, at what we now know to be the peak, and (alas) end, of the Steve Jobs era. The piece opens thus:
Apple doesn’t often fail, and when it does, it isn’t a pretty sight at 1 Infinite Loop. In the summer of 2008, when Apple launched the first version of its iPhone that worked on third-generation mobile networks, it also debuted MobileMe, an e-mail system that was supposed to provide the seamless synchronization features that corporate users love about their BlackBerry smartphones. MobileMe was a dud. Users complained about lost e-mails, and syncing was spotty at best. Though reviewers gushed over the new iPhone, they panned the MobileMe service.
Steve Jobs doesn’t tolerate duds. Shortly after the launch event, he summoned the MobileMe team, gathering them in the Town Hall auditorium in Building 4 of Apple’s campus, the venue the company uses for intimate product unveilings for journalists. According to a participant in the meeting, Jobs walked in, clad in his trademark black mock turtleneck and blue jeans, clasped his hands together, and asked a simple question:
“Can anyone tell me what MobileMe is supposed to do?” Having received a satisfactory answer, he continued, “So why the fuck doesn’t it do that?”
For the next half-hour Jobs berated the group. “You’ve tarnished Apple’s reputation,” he told them. “You should hate each other for having let each other down.” The public humiliation particularly infuriated Jobs. Walt Mossberg, the influential Wall Street Journal gadget columnist, had panned MobileMe. “Mossberg, our friend, is no longer writing good things about us,” Jobs said. On the spot, Jobs named a new executive to run the group.
Tim Cook should have already held a meeting like that to address and rectify this Siri and Apple Intelligence debacle. If such a meeting hasn’t yet occurred or doesn’t happen soon, then, I fear, that’s all she wrote. The ride is over. When mediocrity, excuses, and bullshit take root, they take over. A culture of excellence, accountability, and integrity cannot abide the acceptance of any of those things, and will quickly collapse upon itself with the acceptance of all three. ★
Image Playground would make a ton of sense not as a consumer-facing app, but as an example project for developers. Long ago, Apple used to share the source code for TextEdit as an example project for Mac developers. (TextEdit is actually a low-key great application, though. It’s genuinely useful, reliable, and understandable.) Apple shares tons of sample code at WWDC each year. Image Playground would be a great sample project. The silly app icon even looks like something from a WWDC sample project. What Image Playground is not is a credible useful generative AI tool. Yet Apple keeps talking about it — and showing it off in new hardware demonstrations — like it’s something they should be proud of and that anyone might credibly use for real-world work or even personal purposes. Image Playground does exemplify just how state-of-the-art the generative AI features are in Apple Intelligence, but not in the way Apple seems to think. ↩︎
Skip to the 53-minute mark of Apple’s September “It’s Glowtime” event introducing the iPhones 16, and it’s Craig Federighi who says the following:
“Siri will be able to tap into your personal context to help you in ways that are unique to you. Like pulling up the recommendation for the TV show that your brother sent you last month. And Siri will gain onscreen awareness. So when your friend texts you about a new album, you’ll be able to simply say, ‘Play that.’ And then you’ll be able to take hundreds of new actions in your apps, like updating a friend’s contact card with his new address, or adding a set of photos to a specific album. With Siri’s personal context understanding and action capabilities, you’ll be able to simply say, ‘Send Erica the photos from Saturday’s barbecue’, and Siri will dig up the photos and send them right off.”
That’s about 40 seconds of keynote time I bet Federighi regrets — and that I suspect he was skeptical about including. It’s telling though, that unlike WWDC, Apple didn’t show those features or spend even a full minute talking about at the iPhone 16 event — despite the fact that, ostensibly, those features should have been three months closer to shipping than they were in June. Federighi’s title is SVP of software, and Apple Intelligence and Siri are “software”, but John Giannandrea (SVP of machine learning and AI strategy) is Federighi’s peer, not subordinate, on the org chart — both report directly to Tim Cook — and is responsible for Siri and Apple Intelligence. Why it was Federighi, not Giannandrea, pitching those features in the iPhone 16 event keynote almost certainly comes down to Federighi’s presentation skills and stage presence, not responsibility for the features themselves. But who’s going on camera to pitch these features and promise their future availability the next time? ↩︎︎
Filipe Espósito, reporting last week for 9to5Mac:
As reported by Brazilian newspaper Valor Econômico (via O Globo), a federal judge in Brazil ruled on Wednesday that Apple will have to open up the iOS ecosystem to third-party apps in Brazil just like the company did in the EU. The judge considers that the “limitations” imposed by the company on developers could jeopardize the entry of new competitors in the segment. [...]
But now Judge Pablo Zuniga has ordered that Apple will have to implement the required changes in Brazil within the next three months. The judge states that, despite Apple’s claims, the company “has already complied with similar obligations in other countries, without demonstrating a significant impact or irreparable damage to its business model.”
Presumably Apple will just roll out in Brazil the same compliance rules, policies, and APIs that they started offering last year in the EU. But will Brazilian users get access to EU third-party app marketplaces, and vice-versa? Or will Apple try to segregate these app marketplaces region-by-region, such that Brazilian users will only get access to Brazilian third-party app marketplaces, and EU users will remain limited only to EU third-party app marketplaces?
I asked Riley Testut, the founder of AltStore PAL, if he knew the answers to those questions. His answer, via iMessage:
We haven’t heard anything from Apple (unsurprisingly), but I think it’ll be somewhere in the middle. It’s likely Apple will add some requirements for existing marketplaces to launch in Brazil like they did with EU (e.g. requiring us to have a subsidiary in Brazil) so we’ll all need to get approval again. But assuming marketplaces meet all requirements, I think users in both places will eventually be able use e.g. AltStore PAL and Epic Games Store.
“In the middle” sounds about right from Apple’s perspective too. They’re not going to make this as easy as possible, but I don’t expect them to make it obstinately spiteful either.
Benjamin Mayo 9to5Mac:
Apple is still reeling from the last-week’s news that the most compelling new Apple Intelligence features for Siri have been indefinitely delayed. Over the weekend, it pulled a YouTube ad showcasing personal context running on the iPhone 16. Now, it has updated the Apple website with a new disclaimer wherever the unreleased Siri features are mentioned on the iPhone marketing pages. [...]
The new message to customers found on Apple’s website is different, but equally as vague. It reads:
Siri’s personal context understanding, onscreen awareness, and in-app actions are in development and will be available with a future software update.
Jay Peters, The Verge:
Google has updated its affiliate ads policy for Chrome extensions after creators accused PayPal’s popular Honey browser extension of being a “scam.”
Honey was accused of taking affiliate revenue from the same influencers it paid for promotion by using its Chrome extension to swap in its own affiliate link before you checked out. According to the updated Google policy posted today, this isn’t allowed in most cases
I wouldn’t say they were merely accused — they were caught, red-handed.
Ben Lovejoy, writing at 9to5Mac:
Our editor-in-chief Chance Miller wryly commented that a radical new look would serve as a great way to distract from the ever-slowing progress on the new Siri. But in truth, I think many more Apple users will be wowed by a new look than would ever care about Siri.
If that’s the way it works out — with a new visual look drawing attention from lackluster progress on the AI front — surely the timing will be coincidental, but some accidents are happy accidents, as Bob Ross used to say.
Sure, the new look introduced with iOS 7 was a massively controversial one, and many thought that then Apple design chief Jony Ive should never have been allowed anywhere near the software side of the business. But love it or hate it, you certainly couldn’t ignore it.
The same will be true of a new 3D look, which might even include some (much more modern) skeuomorphic elements. Probably as many will hate it as love it when it’s first introduced, as that seems to be true of any significant change made by the company, but it will likely make more of a bang than any improvement to Siri the company may introduce, then or later.
There should be no question that all of what Lovejoy is saying here is true. If Apple launches an all-new systemwide UI theme for iOS 19, something even half as radical a change as iOS 7’s theme was, it will be the only thing most users notice or opine about. Humans are fundamentally visual creatures and we notice visual changes. And, most humans are resistant to change. A lot of people simply dislike change, even changes for the better. (Honestly, I think that’s what the iOS 18 Photos complaints are about. Complaints about change itself.)
Part of what makes Apple Apple is that the company is (or at least should be) led by people who both have great taste and trust their own instincts. No one’s taste is perfect. Even Steve Jobs pushed through a few clunkers. But if you have great taste and confidence, you’ll do what your gut says is right.
When a company only pushes out changes that avoid controversy, it leads to paralysis. Then stagnation. Earlier today, Cabel Sasser wrote:
One of my strongest early developer memories was being in the “UI Feedback Forum” at WWDC after they introduced Aqua. Think of a live Q&A, but developers giving notes to a team of Apple engineers.
To these veteran Mac coders, the reaction to Aqua was universally negative. People were actively very angry. It’s a waste! It’s ugly! It’s confusing! How could you. It went on and on, and I was surprised because Aqua looked cool and fun to me.
After that WWDC, they never did another Feedback Forum.
It’s true. What a lot of Mac users wanted was a Mac OS X that looked like Mac OS 9.1 The Aqua look and feel was definitely polarizing. And Apple dialed back its most exuberant details with each subsequent Mac OS X update — less transparency, subtler pinstripes (pinstripes!), etc. But iOS 7 was equally polarizing, and its excesses also got dialed back (or perhaps better, said, dialed back up) with each successive iOS release — a little more depth, some subtle hints of texture.
Either Apple is never going to ship an altogether new UI theme, or they’ll ship one and a large number of people will declare it utter garbage and proof that Apple has completely lost its way. Maybe it will be garbage and proof that Apple has lost its way! Or, maybe it will actually prove to be a great new look that starts a decade-long industry-wide trend that all other companies will soon follow — which is what happened with Aqua’s 3D “lickability”, and happened again with iOS 7’s austere flattening. But either way, you won’t be able to judge it by asking for a show of hands from the general population when it’s unveiled. You either have taste or you don’t, and most people don’t, at least when judging something new and unfamiliar. ★
Which is actually what 1999’s Mac OS X Server 1.0 looked like. It really was the best-looking version of the classic Platinum UI theme Apple ever released. ↩︎
Last week The Wall Street Journal laid off about a dozen tech reporters and editors — not on the fun side, where folks like Joanna Stern and Nicole Nguyen work, but on the straight news side. Chris Roush at Talking Biz News obtained a memo sent to the staff from WSJ editor-in-chief Emma Tucker, which concluded:
We’re also creating a new Tech & Media enterprise team, reporting to Sarah, and we’re creating two new jobs on this team based in San Francisco. In the coming weeks, we’ll be reconfiguring beats on the tech team to give more of them a wide ambit. Fewer will focus narrowly on individual companies. We are also advertising for a new tech-focused job on Heard in New York.
These changes do mean that some reporters and editors in San Francisco and New York will be leaving us. I want to thank them for their contributions to the Journal. I want to especially thank Jason Dean, who is departing, for his long and distinguished service.
It was the reporters who focused on specific companies who got laid off, including the Journal’s beat reporter for Apple for the last 6 years, Aaron Tilley. Tilley was last in the news himself a month ago, when one of his former sources, a now-former Apple engineer named Andrew Aude, settled a lawsuit filed by Apple (and who addressed Tilley in text messages as “Homeboy”) and issued the following public apology:
I spent nearly eight years as a software engineer at Apple. During that time, I was given access to sensitive internal Apple information, including what were then unreleased products and features. But instead of keeping this information secret, I made the mistake of sharing this information with journalists who covered the company. I did not realize it at the time, but this turned out to be a profound and expensive mistake. Hundreds of professional relationships I had spent years building were ruined. And my otherwise successful career as a software engineer was derailed, and it will likely be very difficult to rebuild it. Leaking was not worth it. I sincerely apologize to my former colleagues who not only worked tirelessly on projects for Apple, but work hard to keep them secret. They deserved better.
I have no idea whether this fiasco had anything to do with Tilley being laid off (perhaps not, given Tucker’s decision to have fewer reporters on staff who “focus narrowly on individual companies”) but it couldn’t have helped. It’s not exactly a ringing endorsement to have someone publicly declaring that they deeply regret ever having been a source for a Wall Street Journal reporter, and that doing so not only cost them their job but ruined their entire career.
It also probably didn’t help Tilley’s case that he wasn’t very good at covering Apple.
New service for disgruntled/embarrassed Tesla owners:
Our team at Steal My Tesla has years of experience in the automotive industry, with a special focus on luxury vehicles, premium wheels and rims, catalytic converters, and driver assistance systems. Trust us to handle your Tesla acquisition with professionalism and discretion.
A clip from this 2016 interview with Marco Rubio — then a candidate for the Republican presidential primary, today Trump’s secretary of state — by CNN’s Jake Tapper is making the rounds on social media. It’s extraordinary. I’m linking here to the full video, hosted on Rubio’s own YouTube account (for now — watch for this to go down the memory hole) starting at the 7:54 mark:
Tapper: You compared Donald Trump to a third world dictator yesterday in an interview with The New York Times. How so?
Rubio: Well, I don’t know about a dictator. I said a third world strong man. You know, he’s running for president. So no matter what, he won’t be a dictator unless our republic completely crumbles, which I don’t anticipate it will. But, yeah, here’s what happens in many countries around the world: You have a leader that emerges and basically says, Don’t put your faith in yourselves. Don’t put your faith in society. Put your faith in me. I’m a strong leader, and I’m going to make things better — all by myself. This is very typical. You see it in the third world. You see it a lot in Latin America for decades. It’s basically the argument he’s making. That he, single-handedly, is going to turn the country around. We’ve never been that kind of country.
We have a president. The president is an American citizen who serves for a period of time, constrained by the constitution and the powers vested in that office. The president works for the people, not the people for the president. And if you listen to the way he describes himself and what he’s going to do, he’s going to single-handedly do this and do that without regard for whether it’s legal or not.
Look, I think people are going to have to make up their mind. I can tell you this. No matter what happens in this election, for years to come, there are many people on the right, in the media, and voters at large that are going to be having to explain and justify how they fell into this trap of supporting Donald Trump, because this is not going to end well one way or the other. He’s going to be the nominee, and he’s going to lose. Or, he’ll have thrown this party into its most chaotic and divisive period ever. And that’s unfortunate because the Republican Party is the home of the limited government free enterprise movement in America. And if it crumbles or divides or it splits apart, it’ll be very difficult to elect candidates that hold those views at any level of government until we can bring the party back together.
Breathtakingly prescient and succinct.
Andrew Boynton, head of copy at The New Yorker:
Keen-eyed grammar fans may notice some changes in our pages — and in this newsletter. Last fall, David Remnick, the editor, suggested convening a group to talk about the magazine’s house style, to see if any rules might bear reëxamination. The group — comprising copy editors, current and former, and editors — met this past January and came up with a list of styles that might qualify for changes, and in a subsequent meeting the following month the director of copy and production and I came up with a limited list of proposals. It was decided that, while no one wanted to change some of the long-standing “quirky” styles (teen-ager, per cent, etc.), some of newer vintage could go. Along with a few other changes, “in-box” is now “inbox,” “Web site” is now “website,” “Internet” is now “internet,” and “cell phone” is now “cellphone” (though everyone acknowledges that the word “cell” in this context will soon disappear altogether).
Some of you may lament the changes as being radically modern, while others are likely to greet them as long overdue. Welcome to 1995, you may be thinking. (Italicized thoughts are new, too.) Regardless, it should be noted that the diaeresis (see that “reëxamination,” above) has overwhelming support at the magazine, and will remain.
I’m in favor of all these changes except for lowercasing “Internet”. I actually started lowercasing it years ago — I think right around when the AP changed its style — but I was chatting post-Dithering with Ben Thompson a week or two ago and he offhandedly mentioned that lowercasing “internet” is a pet stylistic peeve of his, because there’s only one Internet. He said that and I was like, Yeah! — But so why did I start lowercasing it? The Internet is a lot like the Earth. It’s everywhere. It is our universe, in a sense, from the human-scale perspective. But it’s a unique and distinct thing, thus deserving to be treated as a proper noun. (The universe doesn’t get capitalized because while it’s a one-off, it’s not a name. We speak of the universe like we speak of the planet, which is lowercase.) It’s almost disrespectful to lowercase it, and the Internet is one of the great achievements in the history of mankind.
If these caps were a student project it’d get an F. Who thinks you can print one logo on top of another? They look like mistakes, like caps that got run through the embroidery machine twice. The only good one is for the “ASHOS”, which comes close to the actual word everyone uses for that team of cheaters.
Update, 3pm ET: It looks like maybe MLB and New Era (the hatmakers) have put the whole lineup of caps out of their misery.
Josh Marshall:
I think more than anything I admired Kevin’s restraint and his caution. Blogging is a hustle and the incentives for hyperbole and breathlessness are endless. That makes most people easy to ignore. But Kevin — who had a whole career in the normal-person rat race before he started this — sweated the details. He had a serious mind for facts and numbers and he knew how to work with data. His posts were always overflowing with numbers and charts and levels of detail and nitty gritty I couldn’t pile into my brain because I was too scattered and unfocused. When he said something, you had to take it seriously. When he disagreed with you, you knew it was time to re-check your work. Kevin was almost all signal and very little noise. That was his defining mark.
Marian Drum, posting yesterday on Kevin Drum’s site:
With a heavy heart, I have to tell you that after a long battle with cancer my husband Kevin Drum passed away on Friday, March 7, 2025.
No public memorial services are planned.
In lieu of flowers, please donate to the charity or political cause of your choice.
He was writing right up until the end. No one, on either side of the political spectrum, better wrote about “DEI”. Just in the last week, he had posts on Ukraine/Russia, the LA Times’s stupid new AI-generated “bias meter”, the looming congressional budget crisis, Trump’s tariff crusade. He remained on top of everything that’s going on, not only while cancer was killing him, but while he was stoically keeping his readers fully abreast of his declining health. I mean, Christ, read this update from February 12 — I thought that was it. Dying from cancer is harrowing. He knew the end was near, and as with everything he wrote in his long career, he had the numbers to back it.
I didn’t know Drum personally at all, but there’s a certain kinship amongst bloggers from the very turn of the century. I’ve read Drum’s site continuously for over 20 years and I don’t know anyone whose style and approach can fill the void he leaves. He was obviously a very kind and generous person, but his approach to policy was ruthlessly (and thus to me, admirably) data and evidence driven. Strong opinions, loosely held. He brimmed with curiosity. Who, what, when, and where are all good questions but why is the most interesting one, and Drum always wanted to figure out why. I sure didn’t get his love for cats, but I’m also so glad he shared stuff like that. His approach to actual politics could come across as wonky, but his personality was always there. As natural-born a blogger as anyone who’s ever done it.
I miss his voice already.
Jon Prosser, in a YouTube video with mocked-up animations showing exactly what he’s talking about, 51 days ago:
Today we have your very first exclusive look at the changes coming to iOS 19 — with a redesigned camera app and possibly ... a redesigned iOS.
Basic idea is something very much akin to the look and feel of VisionOS, but brought to the Camera app, and perhaps throughout the entire system (or just parts of it) in iOS 19. Seems cool, seems fresh, and seems aligned with where Apple has been heading.
Here’s Parker Ortolani, himself a talented designer with a particularly keen eye for trends, writing on his blog February 4 after Apple Invites debuted:
The rumored Apple Invites app has arrived. While I obviously haven’t had much time to play with it, I have quickly browsed through the UI and wanted to share some observations. The last new Apple app, Apple Sports, already felt out of place in iOS 18. It has a more visionOS or watchOS-like design language utilizing colorful backgrounds, glassy floating UI elements, expanding buttons, and lots of layered shapes. Apple Invites takes it all even further. It’s got big beautiful cards, translucent cells, big bold buttons, and an emphasis on content. It feels so clearly like a hint of what is to come in a future iOS update. It’s almost screaming it in our faces. It seems awfully suspect that Apple’s two latest apps have a whole new design language that does not mesh with the rest of the OS.
Now comes today, with Mark Gurman dropping this vagueness at Bloomberg:
The revamp — due later this year — will fundamentally change the look of the operating systems and make Apple’s various software platforms more consistent, according to people familiar with the effort. That includes updating the style of icons, menus, apps, windows and system buttons. As part of the push, the company is working to simplify the way users navigate and control their devices, said the people, who asked not to be identified because the project hasn’t been announced. The design is loosely based on the Vision Pro’s software, they said.
That’s it. No screenshots. No mockups. No specific description of the changes. Just “loosely based on the Vision Pro’s software”. Jon Prosser not only said this was coming, but commissioned full, realistic animations to illustrate it, 51 days ago. And astute observers like Ortolani and others have observed, for over a month, that Apple itself is starting to hint at this new design language in its own already-shipping new apps for iOS.
I’m so old I remember when YouTubers made videos about months-old Gurman stories, not the other way around. There’s weak sauce, and then there’s eating paste.
Upgrade:
It’s been a quiet week, so John Gruber briefly joins Jason to discuss Apple’s AI delay, new Macs, new iPads, and the future of Apple regulation worldwide.
Recorded earlier today, so it covers, somehow, all of last week’s Apple news — and last week was a kind of crazy week in Appletown. I even squeezed in some parenting advice for Myke Hurley, who’s out on paternity leave.
Craig Hockenberry returns to the show. Topics include Apple’s new hardware this week — M3 iPad Airs, A16 regular iPads, M4 MacBook Airs, and the M4 Max and surprising M3 Ultra Mac Studios. And we go deep on The Iconfactory’s years-in-the-making new app, Tapestry — a universal timeline for the Internet.
Exclusively sponsored by:
Here’s a statement I got this morning from Apple spokeswoman Jacqueline Roy, verbatim:
“Siri helps our users find what they need and get things done quickly, and in just the past six months, we’ve made Siri more conversational, introduced new features like type to Siri and product knowledge, and added an integration with ChatGPT. We’ve also been working on a more personalized Siri, giving it more awareness of your personal context, as well as the ability to take action for you within and across your apps. It’s going to take us longer than we thought to deliver on these features and we anticipate rolling them out in the coming year.”
This is a Friday sort of wah-waaah sad-trombone news drop, but it’s good that Apple is getting on top of it, and setting expectations accurately.
Reading between the lines, and based on my PhD-level fluency in Cupertino-ese, what Apple is saying here is that these “more personalized Siri” features are being punted from this year’s OS cycle to next year’s: to iOS 19 and MacOS 16. Apple’s years in this context aren’t calendar years, but Apple’s OS product years. Those years effectively start at WWDC. The problem, from Apple’s perspective, messaging-wise, is that they don’t talk about future products. They haven’t officially acknowledged there is going to be an iOS 19 or MacOS 16. They haven’t even yet announced that there’s going to be a WWDC 2025. But they do want to set expectations accurately, especially for a feature as high profile — from Apple’s own marketing push — as Apple Intelligence. So here we are.
It was already pretty obvious these features weren’t coming in iOS 18.4/MacOS 15.4, because they’re not in the developer betas that are already out. And if these features were coming in iOS 18.5/MacOS 15.5, or even iOS 18.6/MacOS 15.6, Apple wouldn’t have felt the need to issue this “It’s going to take us longer than we thought...” statement today. Delivering the “more personalized Siri” in iOS 18.5 or even 18.6 would have been delivering them on the originally announced schedule — the “coming year” that began with WWDC 2024. That would have been the tail end of that “year”, but within it. But Apple, being limited to the self-imposed opaque verbiage of Cupertino-ese, isn’t going to spell that out. But trust me, you can bank on it: what they’re announcing today is that these features are coming in iOS 19/MacOS 16.
Siri has improved in several useful, but iterative, ways under the Apple Intelligence umbrella in this year’s OS cycle (iOS/iPadOS 18 + MacOS 15 Sequoia), but the big “more personalized Siri” that was shown at WWDC last June is the Siri-related Apple Intelligence stuff that people are, justifiably, most excited about. This is the “App Intents” stuff — the features requiring Siri to have access to and, effectively, understanding of your personal information, in your apps, stored privately on your devices. These are the most helpful-sounding, most practical, most futuristic features of Apple Intelligence. And they’re the sort of features Apple is almost uniquely positioned to offer, alongside only Google, as the provider of a mobile platform and device ecosystem.
From Apple’s own announcement, after the WWDC keynote:
Siri will be able to deliver intelligence that’s tailored to the user and their on-device information. For example, a user can say, “Play that podcast that Jamie recommended,” and Siri will locate and play the episode, without the user having to remember whether it was mentioned in a text or an email. Or they could ask, “When is Mom’s flight landing?” and Siri will find the flight details and cross-reference them with real-time flight tracking to give an arrival time.
“When is Mom’s flight landing?” is easily understood, by we humans, and a really good example of the sort of Siri interaction that would be obviously useful. But you can also pretty easily see how complex it is on Apple’s side. Siri has long been able to do the “Mom” thing. Your spouse, your kids — you’ve long able to tell Siri to call or text those people by nickname like that. But getting your mom’s flight information out of your email (which I think was the premise of this demo in the WWDC keynote — that the flight info was stored in a message in Mail) or from text messages, and then knowing how to check for current flight status information for that flight, is something far beyond the ken of current Siri. Getting the information requires access to your most personal information — your own email and messages. Getting the correct information from your messages requires the common sense of understanding that you’re obviously asking about some sort of current trip your mom is taking, not, say, an email she sent you about a flight back in 2019, or one from 2022, or even one from December last year. The Siri we currently know lacks that sort of common sense. And then lastly, after getting the correct flight from your email, performing this task requires Siri to have the agency to go look up the current status of that particular flight — and get it right every time it provides an answer.
If Siri answers, “Your mom’s flight is arriving at 4:30 this afternoon, and it’s currently showing an on-time arrival,” you really want that information to be correct if you’ve promised to pick her up at the airport. You’ll be annoyed if you drive to the airport and it turns out she’s not actually landing until 7:30. And you’ll be in real trouble if she’s landing at 1:30 and you show up three hours later, having banked on Siri’s answer being accurate.
It’s very unusual, to say the least, for Apple to announce a product delay.1 The white iPhone 4 was supposed to debut alongside the black one in June 2010, but, bizarrely, was delayed several times (which I had fun with more than once) and wound up not shipping until the end of April the next year, 10 months late. The original AirPods were announced alongside the iPhones 7 in September 2016, and were supposed to ship in October. They wound up being delayed until December. Apple’s statement in October 2016, regarding the AirPods being delayed, as given to Matthew Panzarino at TechCrunch:
“The early response to AirPods has been incredible. We don’t believe in shipping a product before it’s ready, and we need a little more time before AirPods are ready for our customers.”
This is exactly why Apple hews to an almost religious policy of not announcing products before they’re ready. When they do make exceptions — when they lapse, as it were, as cited above (and below) — they often regret it. You can’t break a promise you never made, and you can’t miss a ship date you never announced. Even with talented teams and realistic managers, there’s often a long span between how long you expect a project to take and how long it actually winds up taking. Engineering schedules often slip even with predictable technologies, and the non-deterministic nature of LLMs makes their future inherently unpredictable.
Most of Apple Intelligence has felt like Apple has pushed it out a year ahead of the company’s usual level of baked-ness. I’ve held all year long that if the entire industry — along with Wall Street — weren’t in the midst of a generative-AI/LLM mania, that “Apple Intelligence” wouldn’t have been announced until this year’s WWDC, not last year’s. Today’s announcement is disappointing, but to me utterly unsurprising. The Siri integration with personal data through App Intents has never felt like something that was going to be ready for real customer usage in this year’s OS cycle. And if it’s Apple’s general policy not to ship a product before it’s ready, that applies tenfold for a product involving LLM access to deeply private on-device information. It’s better to ship late, but ready, than to ship something unfinished or unreliable to hit a promised deadline. And with something like next-generation Siri, I’d say that’s downright essential. Apple can take the bad publicity today’s announcement is engendering; they cannot afford to squander the trust of their users. ★
The granddaddy of modern-era Apple product delays, of course, is AirPower — the multi-device (iPhones, Apple Watch, AirPods) charging device that Apple pre-announced in September 2017, alongside the iPhone X and iPhone 8. When announced, Apple said AirPower would ship in “2018”. It wound up never shipping, but the plug wasn’t officially pulled (so to speak) until 29 March 2019, an absurd 18 months after it was announced. ↩︎
Oliver Darcy, reporting at Status (paywalled, alas), on the sudden demise of FiveThirtyEight:
On Wednesday morning, shortly after sending an all-staff memo announcing layoffs at ABC News, network president Almin Karamehmedovic joined a virtual meeting with the FiveThirtyEight team. The 15 staffers at the political analysis site owned by ABC News had learned the previous evening — via a Wall Street Journal report — that they were being let go. Now, they were hearing it from Karamehmedovic himself.
But Karamehmedovic, like other managers carrying out mass layoffs at ABC News that day, had little to say about it. I’m told he appeared to be reading from a script as he delivered the devastating news. He offered no explanation for why the Disney-owned network had decided to shutter FiveThirtyEight, which it acquired in 2018 along with data whiz and then-editor Nate Silver. He took no questions. He simply thanked the employees for their work, told them a human resources staffer would follow up, and ended the meeting. The entire affair, I’m told, lasted about 15 minutes. Soon after, employees had their access to ABC News’ systems shut off and the FiveThirtyEight website was pulled offline. Just like that, it was over.
Business is business, and layoffs and closures are never anything but traumatic. But there’s a right way to do things and a wrong way. And one of the wrong ways is having the staff find out they’re all fired by reading it from The Wall Street Journal. This guy Karamehmedovic sounds like a real Grade-A dickhead.
And what is the deal with these companies that just unceremoniously pull the plug on websites when they close them? Why not keep the FiveThirtyEight site up and running — at least for a while, if not in perpetuity? It costs practically nothing to run a server for a static/archived website. I don’t get it. It betrays a profound level of disrespect for the work that the site hosted. It’s ABC’s business if they want to close FiveThirtyEight as an ongoing concern, but the years of work deserves to remain online, both out of respect for the people who made it and for the audience that might still want to refer to it. Disney didn’t burn the movies from now-closed subsidiary studios like Touchstone Pictures.
There’s a vast gulf between the old world of print and the new world of online publishing. The difference is profound, but largely overlooked, because it’s human nature to just adapt to the nature of different forms of media and quickly accept current media forms as “normal”. With print, issues were largely ephemeral to the audience. You’d get a newspaper every day, and magazines every week or month, and you’d throw them out after you’d finished reading them. If you wanted to look back at what the paper had published about something a year, a month, or even just a few days ago, you were out of luck, in terms of being able to just look it up on the spot. Librarians did the yeoman’s work of archiving newspapers and magazines to microfilm, so nothing was truly gone or forgotten, but it was an enormous hassle to access microfilm archives. You pretty much had to know the dates of the issues you wanted to see, and you had to do the “searching” with your eyeballs, page by page. The ability of anyone, today, to just instantly access week-, month-, year-, and even decades-old stories is a breathtaking advantage of web publications. And for the last 15 years or so, we’ve had the ability to do this from little machines we all carry in our pockets everywhere we go. It’s truly remarkable.
But the standard behavior when closing a web publication is to just pull the plug. When the whole company goes under, that’s one thing. But when there’s a parent company, especially a thriving one, there’s no justification for pulling the plug other than spiteful disregard for the work. From the perspective of a company the size of Disney, it would cost veritable pennies to keep FiveThirtyEight’s website around forever. What a disgrace. ★