Unauthoritative Pronouncements

Subscribe About

YouTube Stops Screen Stealing On Apple TV

I’m still not entirely sure what happened, but I got a message from Glenn Fleishman the other night asking if tvOS 17.5.1 disabled the YouTube screensaver that I wrote about here and mentioned again over at Six Colors.

The short version is that the YouTube app on Apple TV started showing a slideshow when the TV was paused before the Apple TV’s screensaver kicked on at the default five minute mark. The slideshow from YouTube would function as the screensaver. The slideshow consisted of either the thumbnail of the video you were paused on zooming in, over and over, or a random assortment of still frames from landscape, nature, and drone footage if you happened to be outside of a paused video.

To test what Glenn was reporting I went to my living room Apple TV, which is still on tvOS 17.4. No YouTube screensaver. I asked around a few places and people reported that they don’t have the previously reported screensaver behavior now either.

This isn’t something tvOS did (it obviously wouldn’t be in 17.4) and it didn’t appear to be something YouTube released an updated app for. My best guess is that they have a server-side control for the screensaver behavior and they turned it off there. I don’t have any way to confirm that, but it seems like it’s what fits the facts.

YouTube was very proud of where things were going with screensavers, where they were starting to serve ads in a pilot program. Philipp Schindler, Google’s Chief Business Officer, on the 2024 Q1 investor call:

In Q1, we saw strong traction from the introduction of a Pause Ads pilot on connected TVs, a new non-interruptive ad format that appears when users pause their organic content. Initial results show that Pause Ads are driving strong Brand Lift results and are commanding premium pricing from advertisers.

Fun.

So what happened? What made the screensavers go poof all of a sudden? Is this a temporary reprieve as something about the screensavers gets retooled, or is this permanent? Is it because someone from Apple called someone from Google and got this quietly killed only on Apple TVs?

We Did It, Joe!

I don’t want too jinx it, but I’ve set all my Apple TV screensaver intervals back to the default 5 minutes, instead of the 2 minute workaround I was using before. Hopefully this will never come up again.

2024-05-29 17:30:00

Category: text


Full of Hot Air

A still frame from the BTS 'air head' video. A yellow balloon head with a weird face.
Still frame from the Air Head BTS video. I can see why they didn't go with this generated output.

I’m not a fan of video generators. That’s not to say that I hate “AI” being used in video, it can be a powerful tool, but I hate these things that are not tools.

Sora, and Google’s recently announced Veo, produce RGB pixel output from the videos they were trained on, which seems to be a lot of stock footage.

They don’t make elements, or assets, bits, or pieces, they bake the whole shebang flattened out into a rectangle. If you don’t like that completed video, you can keep on prompting but there’s no file for you to open and tweak.

It’s no surprise that working with the output of these video generators is like working with stock footage. There’s nothing shameful about using stock footage, that’s why it’s there. You can buy explosions, fire, dust hits, blood squirts, wooden debris falling, arc welding sparks, rain, aerial views of cities, etc.

A person has to decide how to combine those elements, which is where Sora and Veo come in. It decides how to combine the stock footage it was trained on, and flatten it out to a new result.

Sometimes the results are uncanny, but they’re only for that one shot. Sometimes the results are warbling bodies, and boiling fingers. A lot of the time, it’s in slow motion.

Air Head

Air Head made by Shy Kids was released a month ago. The people at Shy Kids are clever because they know what would be the most difficult thing to do (having a consistent main character with facial features, expression, and lipsync) so they make the head a featureless balloon. However, they can’t even get Sora to give them the same yellow balloon, attached to the same head, and scaled proportionately to the body the same way. That’s why it’s like a found footage documentary.

Instead, continuity is imposed upon this montage of Sora output. Shy Kids had to slice and dice these individual videos to force even a little consistency in there, even though the only thing credited at the end of the video is Sora.

Here’s a video breakdown released by Shy Kids, on their YouTube channel, not on the OpenAI channel where Air Head is:

“How do you maintain a character, and look consistent, even though Sora is very much a slot machine as to what you get back?” — Walter Woodman Director at Shy Kids

You didn’t, Walter. You cut these shots together back to back that aren’t even close.

Five screenshots from the Air Head short that are shown back to back. There is a male-presenting body seen from the back in each shot and a yellow balloon head. The balloon changes shape, scale, reflectivity, transparency, and how it attaches to the neck.
These shots all cut back-to-back, literally.

Shy Kids followed that first short up with Deflated where they combined live action footage with Sora output and explicitly said that “VFX software” was used.

This looks terrible, on a technical level. If you don’t notice the problems, trust me and watch it again. Look at edges. Look at the neck. Look at the black levels (the darkest darks) and the white levels (check out those posters and billboards).

It also still has the problems from the first video, with the balloon, scale, and attachment point changing wildly throughout, but now there are those matting issues where they’re trying to combine their Sora output with their live action. Sora isn’t a compositing tool, so it can’t insert that footage itself.

In a typical production pipeline you could shoot the reporter and the balloon boy walking down the street together, and then paint out the balloon boy’s head (a relatively trivial matter these days, I assure you). Matchmove the actor’s head and neck, render a yellow balloon asset, and composite it over the painted plate. The balloon would look the same in every single shot and it would match the actor’s performance instead of just wobbling.

They close out the video montage with the reporter teasing an upcoming segment with a doctor. “Why are so many people moving in slow motion? And what’s going on with everyone’s hands?” At least they have some humor about it.

Yesterday, a week after releasing Deflated OpenAI’s YouTube channel released a BTS video from Shy Kids.

They don’t include any botched output from Sora, like they did in the BTS of the first short they released on their own YouTube channel. This time around they show how they stabilize, rotoscope, composite, and color correct the Sora output to fit with the live-action reporter. They also replaced the balloon in a few shots, like the magazine covers.

The most effective uses of Sora in the short are the random montage clips inserted in the “Tough Questions” opener, and the montage of people wearing balloons on a runway. It’s no wonder those are more successful because they’re based on the kinds of input you get from stock footage, and they’re being used in place of stock footage.

What about the aerial shots of the balloon deflating? The first shot worked well, but then they cut to other shots and you could see that it wasn’t the same location, and it wasn’t the same balloon. Sure, they didn’t have to rent a helicopter and gear, but people very rarely do that and instead use … stock footage you can buy online and then put your matching balloon asset on top of.

OpenAI and Google are both selling these video generators as technological breakthroughs in filmmaking. The reality is that it’s artifacting stock footage.

Bad clients, and bad producers will tell their editors to put Sora or Veo output in to the initial edit, then they’ll turn to a VFX house and say that the shots are “90% there” and they “just need someone to take it across the finish line.”

How do I know this? Because that happens with stock footage and weird composites and retimes that editors make in Avid when clients want to have something in the edit so they can figure it out. Even if the client agrees to replace it, they can get married to how the stock footage or temp looked, or how it was timed (remember that playback speed is a factor).

That’s why these companies talk about how it empowers directors, and small-scale productions. Until they want to change something just a little bit.

As a VFX artist I’m all for AI/ML to improve work, but that kind of improvement is in better tracking tools. It’s in better keying and extraction tools. It’s in generative fill for paint purposes. It’s in screen inserts for funky monitors and round rect smart phones. It’s in better photogrammetry with gaussian splatting, radiance fields, etc.

Tools are parts of a pipeline where you can adjust individual elements upstream, or downstream without starting something over from scratch because it isn’t just one flattened hallucination. It’s not a piece of stock footage.

CopyCat and Cattery

Foundry has a tool called CopyCat, where you can train up a model on repetitive tasks, as explained here:

In this example for removing tracking markers from a stack of green paper cards by only painting 5 of the frames to train the model.

Here’s Foundry’s brief explanation of Cattery, a library of models ready to use with CopyCat:

This is how you use machine learning to do filmmaking. These aren’t for making choppy montages.

New breakthroughs in stock footage are being misrepresented by OpenAI, and Google, hungry for investment, and sweaty to establish their product before too many people ask questions. It repulses me as a VFX artist, and it should repulse you too.

2024-05-18 18:20:00

Category: text


Not All Watch Workouts Are Equal

Dr. Drang wrote about how the Apple Watch doesn’t record his location during a paddling Workout, unlike when he uses another type of Workout, like hiking.

It gives you a dot where you started the workout but nothing else. A search of the web to see if I’d done something wrong soon told me that I hadn’t and that lots of paddlers were unhappy about this deficiency.

That’s a bummer. I have no plans to start kayaking, but I’m highlighting this because it seems sort of connected to a pet peeve of mine.

Apple emphasizes when they add Workout types, but they rarely seem to go back and improve ones outside of what really intense fitness people care about, like running and cycling.

I’ll hijack his kayak to complain about how there’s still no detection for resuming Workouts. I walk about an hour in a loop from where I live, and halfway through I stop to get a cold brew at a nice coffee shop. If I forget to start the walking Workout, the Watch will notice, and ask if I would like to begin a Workout, and retroactively note the time I left my house. It’s quite accurate.

However, that coffee shop is sometimes pretty busy, so I need to wait. The Watch will detect the decrease in activity and ask me if I want to pause my workout. The Watch can’t detect if I start walking again. It will stay paused until I remember to unpause it.

This has been like this forever. I have no idea if it is for philosophical reasons, or someone just doesn’t think it’s a problem, but Apple has all the pieces there. Just like how they have all the pieces to record Dr. Drang’s paddling path.


  1. Just as an aside, a not-so-dissimilar organizational problem exists in Settings. Settings frustrates for organization, and for how parts of the Settings panes don’t even function or look the same. A defense of Bad Mac Settings is that people “just search” which is the same dismissal people give for bad file management. However, if you search in Bad Mac Settings for “Desktop & Documents” it returns zero results. If you search for “iCloud” you get several results, and the one you want is “Apps Using iCloud” (because iCloud Drive is an “app”) and then where it says “iCloud Drive” there’s a right-aligned “On >” You might think that just takes you to a toggle to pick between On or Off, since that’s all the nuance they chose to offer, but it opens a modal dialog that grays out the settings (not a navigation to like “>” implies) and then you can see “Desktop & Documents Folders” which has a toggle and explanation. There is a “Sync this Mac” toggle right above that, but that’s for iCloud Drive to have its synced folder on the Mac, it’s not syncing the actual Mac, because why would you do that when you can sync Desktop & Document Folders. But whatever, people just find exactly what they need using search so no need to bother with organization or interface design! 

2024-05-18 12:25:00

Category: text


Apple Still Isn’t Done Building Its Dream iPad ►

Harry McCracken at Fast Company talked to Apple Senior VP of Worldwide Marketing, Greg Joswiak, and VP of Hardware Engineering John Ternus about the new iPads. I wanted to highlight this because it is a rather defensive interview —not defensive toward Harry— but toward criticism of Apple’s iPad line, in particular the iPad Pro, and AI features.

I wrote about the iPad Event from the perspective of it as a marketing event that was a sales pitch not just for new models, but for the platform. If you didn’t think you needed an iPad Pro before the event, I’m not sure why you’d think you needed one after.

Reviews came in, you can go read any of them, and they’re from veteran iPad reviewers who loved the hardware, and leveled the same critiques.

I didn’t feel compelled to write anything about these reviews. As I disclosed before, I’m not a reviewer, or a serious iPad user with serious things to say, but this executive interview bugged me because it rebuffs serious criticism.

iPad Pros are absolutely not bad products, and no one should feel bad about wanting an iPad Pro, buying one with maxed out specs, or only using 1% of its power. Zero judgment.

Reviewers want the devices to have more capabilities that match their hardware, and their ever increasing costs. Which is what makes the interview strange.

“The fact is that the majority of Mac customers have an iPad, and they use them both,” [Joswiak] says. “And a large proportion of iPad customers have a Mac, or even some of them have [Windows] PCs. You use the tool that makes the most sense for you at that time. They’re two different tools.”

They’re two different tools that use the same kinds of processors, the same storage, and the same RAM. The iPad Pros and MacBook Pros cost about the same if you spec them out equally, but what makes them different is mostly the optional Pencil, optional cellular modem, and singular port.

The iPad Pro doesn’t need to run macOS, but the answer to why an iPad Pro can’t do something a Mac can do, shouldn’t be to carry two kinds of computers with the same M-series chips, with the same RAM, with the same storage, and do different things on each.

I see why it’s financially appealing to have two different hardware lines that don’t cannibalize each others’ sales, but that makes the iPad Pro more niche, in a way.

What really bugged me was what John Ternus said about the source of criticism.

But Ternus also pushes back on the notion that the iPad Pro is less than “pro”—a term, he says, that isn’t defined by the Mac.

“There’s a funny perception thing,” he says. “Maybe it’s Mac people with their notion of what professional is. You saw what the Procreate team has done with Apple Pencil Pro. There is no more professional drawing application in the world than Procreate—I mean, they’re the lifeblood of artists.”

Procreate is an exceptional app for illustration. It absolutely deserves all the praise it gets. I’ve enjoyed using it on my own iPad Pro (when I remember to charge it).

It is also the exception that proves the point those “Mac people” are trying to make. That’s one workflow that Apple thoroughly supports on the iPad because of the Pencil, but there is a lack of flexibility for other workflows that don’t need the Pencil, even things as basic as file operations.

Federico Viticci isn’t a Mac person (should have named the site iPadStories amirite?), so it’s worth reflecting on his thorough critique of the platform.

As I noted before, Final Cut Pro for iPad 2 and Logic Pro for iPad 2 seem impressive on the surface, but they don’t handle things like file management and multitasking well. I’ve yet to see a thorough review of Final Cut Pro for iPad like Vjeran Pavic made last year.

Apple didn’t even edit their whole iPad event on an iPad to eat their own dogfood, or describe where and how they had to use the Mac to compliment the iPad as part of that two-device solution.

No one is asking the iPad to do less. No one is trying to look down on anyone that doesn’t want more. There is no zero-sum game where if Jason Snell, Viticci, etc. get what they’re asking for then people currently happy with their iPads will hate their devices.

Circle the wagons, fellas, someone’s complaining they want a more capable $3k iPad Pro!

2024-05-15 21:15:00

Category: text


Gemini In, Gemini Out

This year’s Google I/O event was a strange affair. There was an unhinged DJ who yelled “NO ONE WROTE THIS! GOOGLE WROTE THIS!” while he sort of (?) demoed generative music that he was looping.

Sundar Picahi came out a few minutes later and with the vitality of a mannequin announced that this was “The Gemini Era” and talked about how much progress they’ve made since last Google I/O with Gemini.

Keep in mind that last Google I/O Bard was first made available to everyone. Then Google changed the name of Bard to Gemini this February. They announced an improved version of Gemini 1.5 Pro (a.k.a. Gemini Advanced for some reason?), but didn’t change the version number, as well as Gemini 1.5 Flash, a lighter model, and Gemini Nano which will be embedded in Chrome browsers now, not just Android phones. This is not to get it confused with AI Overlays for Google Search, which can be turned on with the Google Labs flask icon.

The only name Google has left untouched is DeepMind, which is perhaps the most sinister-sounding name possible for LLM and general AI research (Project Astra).

That doesn’t mean that all of this is in anyway sinister, but a lot of it seemed misguided. A lot of it is also very confusing, since there are many Geminis, and they’re going to appear in a variety of places.

There are some demos that everyone in Google’s C-Suite is wild for, regardless of the specific product:

  • Summarizing. Every executive wanted a summary of everything. One summarized an email chain between herself, her husband, and a prospective roofer. The summary said that there was a quote for the work, and the time the work could start but didn’t even include the quote in the summary. She asked a followup question to compare the quotes and that’s when she saw the price. Another exec didn’t have the time to watch a 3 minute video on pickleball rules. Wild that these were selected as demos.
  • Meal planning. We saw two sets of meal planning examples in the presentation. It showed off how you could load up a prompt (a question) with terms and then you’d get back breakfast, lunch, and dinner recipes. Individual UI elements existed to override a particular item, so it wasn’t like you were locked in, but these weren’t really any different from the recipes you’d get doing a Google search before this rolls out. It wasn’t writing a recipe, showing the recipe, doing measurement calculations or generating a shopping list. These are links to all the recipe sites that are laden with shady ad-tech cruft and SEO keyword stuffing to try and get into Google search results. I wasn’t as wowed as these busy professionals.

These are dreadful things to watch, and are not really as impressive as executives seem to think that they are. I hope that Apple doesn’t fall into this trap at WWDC.

There was only one travel planning demo, so I didn’t include it above, but it was a lengthy one. The exec had already booked flights, and a hotel, and that information was in her Gmail. She constructed a prompt to get help organizing what to do and where to eat according to that flight and hotel information. The results were produced and she could browse and override individual bits, but budget and prices really didn’t seem to factor in. These restaurants are also things you could just … Google for instead of paying $19.99 a month for Gemini Advanced. Who’s that stressed about planning they’re paying that fee?

Surely, at some point that might filter down to regular Google Search, but maybe Google is planning on Gemini being so exciting that people start paying for it?

There were some good demos about being able to load up a bunch of documents and pick out important information from them. More than just opening each and performing a text search. Also that data is explicitly not used for training models, and Google doesn’t use it. That sort of thing could have interesting applications.

I was a lot less happy with the demonstration of a virtual teammate that sits in Google Workspace. In this case, named Chip. The first hypothetical scenario that the presenter invents for Chip is to “quickly catchup” by asking the Google Chat space, “[Does] Anyone know if our IO storyboards are approved?”

If anyone asked the group that general question, spamming everyone, he should have read the channel updates first or done a search for “storyboards” maybe check in with the person responsible for approving them. Instead, everyone gets spammed and then gets spammed by Chip’s reply, which is, “Based on the Google I/O chat between Aparna and Kristina it looks like the storyboard is approved”. Yeah, for some reason it doesn’t use punctuation to appear more human-like. Also, it couches it’s response with “it looks like” to seemingly avoid legal liability? Remember, Gemini, like all LLMs isn’t a reliable source of truth.

Congratulations, you spammed everyone in the chat so you look like a fool, got a bot that replied without any certainty, and still should check on the approval state. If those storyboards weren’t approved you’d be in a position of trying to tell them this was Chip’s fault.

Then he follows that up by demoing Chip summarizing where they’re at on their schedule, and it highlights a potential conflict. Another person offscreen asks for a summary.

These are not tasks that require automation, because you should have hired capable people. We should appreciate labor that goes into all aspects of communication and not treat our conversations with one another like a free-flowing firehose.

What is not demoed, and what I’m sure will appeal to bad bosses around the world, is the capacity to use this tool to micromanage employees, or generally snoop on progress in an invasive and disrespectful way. Chip doesn’t care about summarizing your status for that boss, or making any mistakes, because Chip isn’t a person.

Creativity

A constant source of tension with generative AI is over training sources, and whether the application is a tool, or a replacement for an artist. Google is not transparent about the datasets it trains on, so we’ll just take it as a given that there’s stuff in that training data that people would object to.

Setting that aside, we started the I/O event with the guy using Google to make a short clip of nonsensical music that he looped. That part was very much not using Google’s tool. It just generated that little snippet and that was it.

Doug Eck came out on stage later in the presentation to talk about Generative Media - image, music, and video.

Imagen 3

more photo-real, fewer distortions and artifacts, better text rendering, and “independent evaluators preferred Imagen 3 over other popular image generation models.” It really doesn’t seem all that distinct in the demo, and I am definitely not the target audience for this. There’s little an artist can do with the output, so this continues to be mostly for someone that couldn’t produce artwork.

Music AI Sandbox

Creates instrumental sections “from scratch” and transfer “styles” between tracks. Wycleaf Jean appears in a video to describe how he considers the tool to be akin to sampling. “As a hip hop producer, we dug in the crates. We playin’ these vinyls and the part where there’s no vocals, we pull it, we sample, and we create an entire song around that. So right now we’re diggin’ in the infinite crate. It’s endless.”

Then my nemesis Marc Rebelliet appears and talks about how he uses it to generate a bunch of loops. “Google’s loops right here. These are Gloops.”

Sigh.

Veo

“High quality” 1080p videos from text, image, and video prompts. One of the demo started from a video, and extended it. Then to show us what it can really do they put it in Donald Glover’s hands. Cut to Donald Glover saying he’s interested in AI. Then there are a lot of vague clips of things where you can see some warbling, and the ground surface artifacting like crazy with the cowboy boots. That’s it though, they didn’t actually have the short film they allegedly were making with Donald Glover.

Veo will apparently only be available to select creators at lab.google and there’s a waitlist open now. But… what does it do? How can you edit or adjust the output? Can someone fix those cowboy boots? Can someone keep any kind of consistency from shot to shot so it doesn’t look like it’s completely different each time you generate a video? How are you going to handle generating sound to match with the video you’re generating?

Update: The videos have a maximum limit of 60 seconds. Good grief.

I’m the most skeptical of generative video at the end of the day. These things approximate stock footage —probably because they used a lot of stock footage in their training data? Possibly. There are some more videos on their labs site so you can see things tearing and burbling.

I don’t think it is responsible for Google, or OpenAI for that matter, to sell fully generative video as being something that’s right around the corner.

Not a lot of producers are technically savvy, they’ll believe this stuff, and it’ll make a big mess.

In Summary

I think this was a cynical event trying to apply AI to things as fast as they can get it out the door. Building a business model on the fly to charge for computer resources. Inculcating LLMs into things that are not always improved by having them there. Impressing the inveterate gamblers of Wall Street to show that you have “AI” like OpenAI does.

There’s intriguing stuff in here, to be sure, like the Astra demo, and checking through your personal files with a level of context awareness that a search lacks.

But summarizing? Meal planning? Increasing office dysfunction? Suspicious generative video?

Sundar even made a heavily scripted, cringeworthy joke out of it at the end of the presentation where he mentioned someone was probably counting how many times they said “AI” in the presentation. Then the script text file (not even the video output up to that point) went into a prompt and a Gemini model counted 120 times. Was that even correct?

I know it’s to show off feeding data to the model and asking it to do something, but it’s an oddly accurate metaphor for this presentation where Gemini didn’t really need to be used, and it didn’t really improve anything.


  1. Just as an aside, a not-so-dissimilar organizational problem exists in Settings. Settings frustrates for organization, and for how parts of the Settings panes don’t even function or look the same. A defense of Bad Mac Settings is that people “just search” which is the same dismissal people give for bad file management. However, if you search in Bad Mac Settings for “Desktop & Documents” it returns zero results. If you search for “iCloud” you get several results, and the one you want is “Apps Using iCloud” (because iCloud Drive is an “app”) and then where it says “iCloud Drive” there’s a right-aligned “On >” You might think that just takes you to a toggle to pick between On or Off, since that’s all the nuance they chose to offer, but it opens a modal dialog that grays out the settings (not a navigation to like “>” implies) and then you can see “Desktop & Documents Folders” which has a toggle and explanation. There is a “Sync this Mac” toggle right above that, but that’s for iCloud Drive to have its synced folder on the Mac, it’s not syncing the actual Mac, because why would you do that when you can sync Desktop & Document Folders. But whatever, people just find exactly what they need using search so no need to bother with organization or interface design! 

2024-05-14 17:00:00

Category: text


The iPad Event

I was struggling to make it through yesterday’s iPad event video. At one point I paused and went outside to do some weeding. That’s how captivating the event video was. Weeding!

Part of that is the fault of Apple’s formulaic and sterile presentations, which are not a new phenomenon at this point. Lex Friedman, and others, would like Apple to bring back live events to get some life back into these things. I doubt they’d give over this level of control for the chaos of live events, but just something recorded that’s more energetic. It’s giving “high school presentation”. It’s giving somnambulant, honey.

The other issue was the subject matter —iPads.

As I said on Mastodon:

If you are someone who regularly uses an iPad, and you needed new hardware for some reason, then any new iPad hardware is an iPad for you. If you didn’t use an iPad (or had one collecting dust on a shelf) I don’t know why today’s announcements would make you want to buy an iPad.

That’s what I keep coming back to when I consider a media event like this. This was a big production, both in terms of the video itself, and the dual press events in New York and London. It’s not nothin’ to go through this effort to pitch these iPads to consumers either directly, through the sleepy video, or indirectly, through the press.

To go through all that effort and the appeal of the new iPad Air is that it’s like an older iPad Pro, and that the iPad Pro is a thinner iPad Pro, is … well … underwhelming if the hardware wasn’t a primary concern for you before yesterday.

The tandem OLED is great. The M4 sounds pretty amazing (I can’t wait until that display controller finds its way into Macs). The Pencil Pro seems nice, if perhaps a little over engineered for people used to Wacom tablets that have been asking for physical buttons and comfortable grips. The new Magic Keyboard seems very MacBook-like with its function row, and big touch pad.

However the headline feature that Apple thinks will knock our socks off is that the iPad Pro is Apple’s thinnest device ever. [Sound of crickets chirping.]

This is an especially challenging sales pitch when the price of an iPad Pro has ratcheted up a little, and it needs new accessories if you want to do those fancy things. You wind up spending more than it would cost to buy some Macs. Yet, if you spent that comparable sum, you might have a far less capable machine because of the tremendous peaks and valleys in what an iPad can do.

The consistent refrain before, and after the event is that Apple isn’t addressing the iPad software platform.

Jason Snell:

What I’m saying is, when it comes to iPad Pro hardware, it feels almost like Apple can do no wrong. On the software side, iPadOS is still rife with limitations that probably don’t matter much if you’re just using it to watch TV in bed or triage a few emails—but matter a lot if you’re trying to go beyond a limited set of features and some specific apps.

I will live in hope that the next version of iPadOS will address some more of these issues. (I have expressed this sentiment every single time a new iPad Pro has been released. It hasn’t helped.)

Federico Viticci:

The elephant in the room, which I plan to address more in-depth in a future story, is that while the iPad Pro’s hardware was fine before and is finer now, its software was a letdown before, and nothing has changed after today’s event. I don’t need to rehashwhy I think Apple is missing a huge opportunity by not embracing the iPad Pro as a machine that could do both iPadOS and macOS equally well in the same package. I’ll save that for iPadOS 18 at WWDC. What I will say is that if there was a gap between the older-generation iPad Pro hardware and its software, that gap is now a Tears of the Kingdom-sized chasm between these thin, OLED iPad Pros and iPadOS 17.

Marques Brownlee:

“But the thing is, and I feel like we’ve been saying this for years, is it kind of doesn’t matter how powerful they make the iPad. It’s still an iPad, right? It’s still iPadOS. And we’ve seen gigantic improvements in the M-series chips. And these iPads are like the most powerful chips on paper ever, but they’re still iPads. So the last thing we need after all this time is just another spec-bumped iPad, right? […] So in this awkward meantime [between iPads shipping next week and WWDC], here we have these really, really impressive spec-bumped iPad Pros, but the list of things it can do is the same as my three year-old M1. Just saying. What a time to be alive.

While Stephen Hackett didn’t attend the press events in person, he’s got a pretty succinct critique on his blog:

As nice as the new OLED display looks, and no matter how powerful the new M4 may be, the iPad’s problem in 2024 — or another year for that matter — is the software. Fourteen years into its lifespan and the iPad still can’t seem to fully shake off its iPhone OS roots. Almost everything Apple has attempted to bolt atop iPadOS to make it more useful for more people has come with weird tradeoffs. Look no further than something like Stage Manager, or that just today Apple announced a version of Final Cut that can use external drives for project storage.

So, like I was saying, there’s no sales pitch here for people that were previously uninterested in iPads. As if the maximum addressable iPad market has been reached and now the only way to move the needle on sales is to entice existing owners to upgrade.

The Mac is still pitted against the array of PC vendors out there, so it does have a sales pitch to those PC buyers and isn’t just reliant on its own iterations. The iPad is also poised as a PC replacement, but it’s always depicted as a more appliance-like replacement.

Send some emails, use QuickBooks on BART, “catch up on a hit show, like Palm Royale.”

People do use and love their iPads, so perhaps what it is is enough. It serves a role.

The iPad might not ever need to be more than the iPad is now, but at this point you know if that aligns with you or not. Unlike some others, I’m not expecting any dramatic innovations at WWDC this Summer, and even if there were you’d be on a beta iPadOS until the Fall if you really wanted to use them.

But the Pro Apps

I am really at a loss when it comes to Apple’s Final Cut Pro for iPad 2 and Logic Pro for iPad 2. Lovely names. I haven’t heard of anyone using the first versions (which is not to say that no one uses them, just that if there was a professional non-linear editor for $4.99 a month you’d maybe hear about someone using it.)

Apple might have thought it sounded impressive when they punctuated the event with, “Shot on iPhone. Edited on Mac and iPad.” That’s not quite as impressive as the fall MacBook event that was shot on iPhones. I’ll be interested to see if they release a BTS video in a few days that shows us how much of this was Final Cut Pro for iPad. At what point did they export the project files on that one-way trip to the Mac? How much did they render on the iPad?

Functionally, they still don’t match the desktop counterparts feature for feature. Like exporting video, which you can’t do in the background or it will kill the export.

The Final Cut Pro for iPad project file format continues to be incapable of round-tripping between a Mac and back to an iPad. It’s a one-way trip.

As already noted, the project files can at least finally live somewhere other than the iPad’s on-device storage. I’d love to hear an explanation about why that feature took this long.

They still can’t use disparate files in the file system though, which is bananas. Sure, you reduce the chance that someone will open a project file to find missing media, but you also bloat this opaque package file container, and need to pay attention to whether or not you have “Include All Media” checked when you export your project for a Mac or you lose anything that’s not being currently used on the timeline.

I do understand that things are this way because iPadOS file management is based on iOS file management, and that can’t ever be as complicated as a Mac because it might hurt people’s wittew bwains, but aren’t these pro apps supposedly for people that would use Final Cut Pro and Logic Pro on a Mac? Who is the target market?

Personally, I was also a little let down that the new features for announced for Final Cut Pro for iPad 2 and Logic Pro for iPad 2 were mentioned to bolster the iPad, when those same features were coming to the Mac. Not because I wanted those features to be exclusive, but because it felt misleading to frame them as iPad features and quietly mention the Mac.

With the notable exception of Final Cut Camera (woof, what a name) multi-cam support, which apparently a Mac can’t handle. It must be because the file system is too complicated on the Mac.

Back to the Future

The main tension here seems to be people who want to be able to use an iPad as a complete drop-in replacement (or merely on-the-road substitution) for a Mac in as close to 100% of the circumstances that they would use a Mac for. Otherwise that is all that M4 horsepower for? Jason and Federico have both opined that the solution to this ought to be with Mac virtualization. Letting people choose to run a macOS environment on their iPad which has the same M-class chips.

If people are asking themselves “how much RAM should I get for my iPad?” Then maybe we’ve crossed into more Mac-like territory than people are willing to admit.

I don’t think that is an unreasonable request, and it seems to be the simplest route to appease those users, while also leaving the basic iPad experience unmolested.

Fold iPadOS back into iOS. The jig is up anyway. When iPadOS was split off from iOS it was supposedly to let it be its own thing, but that hasn’t happened. It’s just a platform Apple can deprioritize when they’re focusing on getting iOS out for the far more important new iPhones.

Let iOS be Apple’s friendly, touch operating system. Let macOS be Apple’s slightly less-friendly OS for power users.

Would it be a great touch-first experience to use today’s macOS on the iPad? No, but that’s way less of a problem than some “but but” nerds make it out to be, because of accessories, like the new Magic Keyboard that looks just like the lower half of a MacBook Air, touch pad and all. Universal Control, virtual desktops, etc. People are already capable of using that UI on those devices. No one would be required to use it anyway.

In my humble opinion, it seems much more difficult, and fraught to revise everything from the file system up so that not everything needs to live in package files. To allow background export and rendering entitlements so people could actually multitask. An honest to god Terminal app where people could install things like Python, or Node to do development (even if that was sandboxed from the system processes, but could mingle freely with files).

Anyway, this is my two cents, as someone that can’t remember the last time he charged his iPad Pro. Make of it what you will, but definitely listen to the far more exuberant iPad users that feel a little bummed out by the best iPad Airs and the best iPad Pros ever made.

2024-05-08 17:30:00

Category: text


Prime Video Steals the Show

In the continuing quest to suck the lifeblood out of us all, Amazon announced three new streaming ad formats for Prime Video (well, 2.5). This is in addition to all the other stuff they were doing with ads in the interface and screen stealers “pause ads”. Scharon Harding at Ars Technica first brought this to my attention, but it’s worth reading Amazon’s advertising blog post about this:

  • Shoppable carousel ads, which make it easy for customers to browse and shop multiple related products on Amazon during ad breaks on Prime Video. Brands can present a sliding lineup of their products that customers can explore on Amazon and add to their cart using most living-room remotes. The ad automatically pauses so that customers can browse, and automatically resumes play when ad interaction has stopped.
  • Interactive pause ads, which enable customers to discover and engage with brands when they decide to pause the show or movie they’re streaming. When viewers press pause on their living-room remote, they will see a translucent ad featuring brand messaging and imagery, along with an “Add to Cart” and “Learn More” creative overlay. These ads extend the engagement opportunity beyond a traditional ad break, as the interactive overlay is available to customers for as long as the content is paused. With a click of their remote, customers can easily add the product to their Amazon cart, get more information sent to their email, and resume their stream at any time.
  • Interactive brand trivia ads, which help advertisers elevate their storytelling by entertaining customers with factoids about their brand while giving them the opportunity to shop on Amazon, learn more about services and products, and even unlock rewards. Customers can use their living-room remote to add a product to their cart, request information via email, and claim rewards like Amazon shopping credits with the purchase of eligible items.

Guh-ross.

Let’s just read some more about this depredation, shall we?

Prime Video has an average monthly ad-supported reach of over 200 million global customers. With Amazon customers shopping while watching content on Prime Video, Amazon Ads connects content to customers using Amazon’s addressable signals and first-party audiences. With this set of innovative S TV ad formats and access to a closed loop of insights, billions of signals help brands to continually improve their ad performance and campaign strategy.

Translation: We have engineered a captive audience by flipping that switch on opting everyone into ad-supported plans, and we have all the data on the audience from those plans, and if you want access to them, and the “insights” from their data you’ll use our advertising platform.

Let’s hop back over to Ars, where Scharon points out:

Still, Amazon claimed today that Prime Video ads reach an average of 200 million people monthly. Although, Amazon hasn’t provided a firm figure on how many Prime Video subscribers it currently has overall. In 2021, Amazon said that Prime, which includes Prime Video, had 200 million subscribers.

So that 200 million number is a lie because not every Prime subscriber watches Prime Video. They have the capacity to show ads to 200 million subscribers, were they all to actually use Prime Video.

This offends me for the same reason as all the other stuff, not because of advertising in the abstract, but because it is worsening an experience in a way no one anticipated when they subscribed. It is altering the deal when all this time I’ve been praying they do not alter it any further.

I do wonder how long it’ll be before they start offering advertisers display banners framing the video content. You know, but in a tasteful way that respects the closed-loop insights.


  1. I hope no one said that, “When God closes a door, he opens a window” line because that is just the worst. “This could be a good thing for you.” Is my runner-up. 

2024-05-07 17:00:00

Category: text


Simply Make More Vision Pro Videos

One of the consistent laments I’ve heard on podcasts over the last couple of months is that there’s not enough “content” specifically for the Vision Pro.

People have pointed towards Apple’s sports push, with live sports offering an opportunity for Immersive Video. All Apple’s put out is a widely-panned, five-minute highlight reel.

It kind of proves that this stuff just really isn’t that easy, or inexpensive, to make. I doubt that anyone at Apple is surprised about the return on their investment.

What little catalog content that exists is mostly a byproduct of 3D movie distribution, which is why it’s the most plentiful source of Vision Pro media. That’s more about negotiating deals, old-school iTunes style, for stuff that has no real home video venue (practically speaking, stereoscopic home video didn’t take off).

So let’s assess the video sources:

Old School

2D Movies and TV

The sales pitch of a huge screen (that feels like it’s only a few feet in front of you) hasn’t proven to be all that attractive based on my anecdotal observations. The supply of 2D Film and TV is relatively unconstrained except for the absence of Netflix and YouTube. It’s not especially enticing compared to other devices.

2D Live Sports

The main benefit for this 2D experience is multi-window viewing for watching several games, or stats, at once. Like a sports bar strapped to your face. I haven’t observed a lot of chatter about it except when new app releases with new viewing modes become available. I’m unsure if people are really continuing to watch games in the Vision Pro after they write up the app updates, or if the pull of screens further from your eyeballs wins out when there’s nothing new.

3D Movies

This is the byproduct of theatrical 3D. It’s still a windowed experience on a big (close-feeling) screen where things mostly recede into the screen, for reasons I have discussed before. I’ve seen some people respond favorably to a handful of viewing experiences, but the limiting factor seems to be the same as 3D theatrical —is there a reason to see this in 3D at all? Or can I just watch it 2D with no downsides?

Remember that Apple has no 3D movies on Apple TV+. Not a thing. They had no reason to have the theatrical stereoscopic byproduct that legacy studios make, so they can’t just drag those files over. Their movies, at present, don’t stand to benefit from stereo-conversion either because unlike Apple TV+ shows, the movies they greenlight aren’t blockbuster spectacles, they’re targeted mainly at securing prestigious awards for talky-dramas.

3D TV Shows

What? LOL. Get outta here!

This isn’t really a thing, because 3D never took off for home viewing in the 2010s, and had no theatrical demand to prop it up. There is no catalog like there is for 3D movies.

However, that doesn’t have to be true. Apple is in an interesting position right now where they need compelling material for the Vision Pro, they have a lot of 2D TV shows they make themselves that have spectacle for 3D, and a VFX market that is almost completely dead because of reduced production following the 2023 strikes, and the potential 2024 IATSE strike. Additionally, Apple TV+ series have short production runs, and few seasons. It is kind of the perfect time to do stereo conversion on some of their popular shows like Foundation, Monarch, or For All Mankind. Where possible the studios that did the original work for CGI scenes could even be contracted for stereo renders because they weren’t produced aeons ago.

Would that make headsets fly off the shelves? Probably not, but it builds a catalog Apple simply doesn’t have from themselves or others. Compare that to live events, where there’s little interest in watching old events, but there is definitely interest in rewatching TV series. Remember only a teeny tiny percentage of people have seen these shows at all to begin with, so it’s also an opportunity for new customers.

Then when new seasons are available they could be available in 3D. Someone could catch up on For All Mankind and spring right in to new, weekly stereo releases.

If they want to create a market, this strikes me as a not-ridiculous candidate.

Immersive Video

While people talk about Immersive Video as one thing, I think it’s really worth discussing it based on subject matter, because I think that makes far more of a difference than the technical details of stitched together spherical projections.

Live Sports Immersive Video

This is the one I’ve seen and heard the most requests for. People keep saying you can just stick some white-obelisk camera rigs in stadiums and that’s it.

I’ve never done live TV work, which is a very different pipeline from my job, but it seems that you still have to solve for graphics, and editorial choices that have different constraints from 2D or even 3D. You also have equipment and crew dedicated to this because it is not piggybacking on what the 2D broadcast team is doing. It’s a different medium.

The most obvious evidence of this not being simple as “just” putting camera rigs there, is that if it was that simple, then we wouldn’t be talking about any of this.

The sporting events are also at the mercy of the calendar, what deals have been cut, etc. They have far less value as catalog content (though there are people that watch old games, I don’t think a game from a couple months ago is “old” enough).

Live sports are definitely a way to increase adoption of a technology. Not everyone bought an HD TV when they hit the market, but someone knew a person that had one, and that person might show “The Big Game” or whatever and then when the next person in that friend group could buy a TV they’d buy one, and so on.

The exception, of course, being 3D live sports. Not many people remember ESPN 3D, but let us take a moment to ruminate on it’s lifecycle from it’s 2010 launch with a smattering of 3D broadcasts, to it’s 2013 demise for being too expensive to produce and having low consumer demand.

I’m sure none of that history is relevant to this discussion at all.

Immersive Video Experiences

There are still a handful of these. Whether or not you like them depends deeply on the subject matter. I was not particularly enraptured during my own demo experience, but that’s not saying much at all. I freely admit that I’m a tough audience.

I’ve heard Casey Liss groan about the dinosaur thing not really being for him. Just like with any other movie or TV show, people have to want to watch it otherwise it might as well not exist. Which is why there needs to be both a high quantity and a high variety.

That’s no simple task. Even though the current immersive video experiences were undoubtedly expensive to produce (especially relative to 2D) it’s going to take even more that what they’re spending, or just a very, very, very, long time to accumulate a catalog.

The problem with slowly producing material on an infinite time scale, is that while Apple certainly has the money to whittle away at this, they don’t have infinite consumer interest over that same period of time.

If the rumors are true and the next headset hardware is going to be released in 2026, then we’re looking at probably 3x what the current library is if they maintain output at this rate. I’m skeptical that will constitute a substantially more impressive catalog.

User-Generated Videos

Apple’s really expecting a lot of people to be interested in shooting Spatial Video, which is just a fancy term for 3D video with blurry edges.

Spatial Home Videos

Spatial Videos don’t have to meet the same standard of quality as studio fare (they also quite literally can’t.) However, Apple has no venue for hosting user-generated video to share with wide audiences so these videos are only hyper-relevant to individual people and entirely reliant on the person choosing to make the videos for themselves.

The upcoming iPhone 16 has their cameras moved around so it’s supposed to be able to shoot Spatial Video too, but it remains to be seen what quality improvements there will be. Sacrificing all the cool stuff that your iPhone can do in 2D, is incredibly unfortunate at present.

It’s important to remember that adding Spatial Video to the regular iPhone 16 doesn’t mean much in terms of Vision Pro adoption, because the people buying iPhone 16s are likely price sensitive. That’s why they’re not buying Pros. So good luck selling them a $3500 headset to look at their own home movies in subpar quality.

VR/3D User-Generated Video

Let’s circle back to YouTube, which is absent from the Vision Pro because of pride, business, and all that jazz. People have been uploading 3D videos to YouTube for over a decade, and a VR “immersive” videos since 2015 that are all available on Meta Quests, and whatever monstrosities wind up shipping with Horizon OS.

I’m not going to debate the quality of user-generated video in the abstract, but I will note that the low production values of YouTube creators in any dimension don’t seem to be a huge issue.

When something’s not from a studio, they don’t have the same expectations, or the same para-social relationships for that matter.

Apple should leverage YouTube’s 3D and VR videos here if for no other reason than to have the minimum level of content available to people with headsets.

If You Build It

The problem with Apple’s strategy from the outset is that it relies on the assumption that people will beat down the doors to make something for Apple’s platform, and Apple can collect money from movie rentals, sales, and subscriptions. They’ve got demos.

It’s really on Apple to invest an ungodly amount of money instead of waiting for an external spark of interest from some other studio. There’s no market to entice anything tailored to it, specifically.

Despite Apple and Disney’s special relationship which produced the Disney+ Vision Pro app, Disney’s not going to throw money at a Vision Pro app forever for Apple. The clock is ticking.

Disney didn’t throw money forever at ESPN 3D, but that was waaaaay back when the CEO was Bob Iger so I’m sure it’s different now (checks notes).

In all seriousness, they should have spent the money they didn’t want to spend before. Now they’ll have to spend even more, and they have less time to do it in.

Knowing they don’t have a lot of time I think stereo conversion of their hit action shows is the fastest turnaround. With new episodes also coming out in 3D. That’s not instant, but it’s a lot faster, and more durable than the live sports they’re fumbling.

No external entities are invested in the success of the Vision Pro hardware or visionOS software — unless Apple pays them to be.


  1. I didn’t read the fine print, but I’m pretty sure there’s a complex system of licensing arrangements that keep it from being “all mine”. 

2024-05-06 15:00:00

Category: text


Florals? For Spring? Groundbreaking

Three, large, purple orchid flowers.

This is a pretty good time of the year to go outside and appreciate what’s happening in nature. Even if you’re in a major metropolitan city, like me, you’ve still got flowers out the wazoo. Knock back a few antihistamines, grab a camera, on go on a little walk.

A yellow venus slipper flower with wine-colored dots.

It could be the camera on your smartphone, but try to grab something you don’t ordinarily take pictures with just so you stretch some different mental muscles.

Shockingly, almost exactly a year ago, I wrote up a post for Six Colors about this topic.

You’re unlikely to win any awards, since the subject matter isn’t all that special, but it’s for you more than it is for other people. The act of composing a shot, adjusting your aperture settings, selecting your focus, etc. can clear your head.

The other week I grabbed my cameras and went to The Huntington Library and Botanical Gardens in San Marino (the Beverly Hills of the San Gabriel Valley).

The Chinese Garden at the Huntington on a sunny day.

I have a Nikon D80 that I bought used last year, and a used Sigma 17-70 F2.8-4 lens. It is absolutely not what I would take with me if I was worried about flexibility, dynamic range, low-light, lightness, space, etc.

The bright yellow inflorescence of an aloe vera plant with a yellow bee flying toward it.

It’s for jaunts. For outings. The thing is a brick that takes lovely 10.2 megapixel photos.

The desert garden section of The Huntington Library full of cacti and succluents with colorful flowers and a couple towering palms.
The Desert Garden at the Huntington in Spring.

It doesn’t have to be a botanical garden though. There’s plenty of stuff in the most mundane of urban locations.

Orchid tree with pink, purple and white flowers.
Bauhinia variegata, a.k.a. orchid tree. These are blooming all along the streets.

I just went on another little walk around on a nice, sunny, Sunday morning, with the ~lead weight~ D80 and it felt good. None of the photos are impressive or precious memories, but it felt good to do it.

A male house sparrow on the branch of a hibuscus plant.
Holly growing through a chain link fence.

Again, there’s nothing about these photos that’s exciting or novel. The old Nikon D80 CCD sensor isn’t magic. This subject matter isn’t special. These photos aren’t groundbreaking.

You should also go out with a hunk of junk and make your own achingly basic flower photos too.

The purple inflorescene of a Pride of Madeira plant with small, yellow bee.

  1. I hope no one said that, “When God closes a door, he opens a window” line because that is just the worst. “This could be a good thing for you.” Is my runner-up. 

2024-05-05 17:00:00

Category: text


Logitech’s Mouse Software Now Includes ChatGPT Support, Adds Janky ‘ai_overlay_tmp’ Directory to Users’ Home Folders ►

Stephen Hackett noted that Logitech has added some extra bullshit to their bullshit Logi Options+ app in a post you absolutely should read.

I cannot tell how little I want THE SOFTWARE FOR MY MOUSE to include features tied to ChatGPT … let alone a mouse with a built-in button to start a prompt.

You need this garbage app to take full advantage of your Logitech hardware. It’s a shame because the MX Master 3S is an excellent mouse. My favorite mouse ever.

What makes it my favorite mouse ever is the thumb-button and its accompanying gestures. Click that paddle-ish button once and you get Mission Control. Hold it down and swipe left and it swipes to the Space “to the left” the same goes for the right.

screenshot of the Logi Options+ app showing the gesture controls.

Most people don’t think they use Spaces, but every full-screen app is a Space. It’s true. I deal with a lot of full screen apps. My employer used remote desktop software, and the best way to use that is full screen.

It’s pretty indispensable to be able to hold and drag to pop from app to desktop to app.

I might have gone for a couple months without even noticing the AI cruft, because I don’t launch Logi Options+, but I would eventually have noticed the folder, or the running process, like Stephen did and waste my time trying to figure it out.

In theory it’s not hurting anything because it’s not doing anything, but in principle my computer is not at Logitech’s disposal. Much like my recent complaints about YouTube, or all the other companies, my devices are mine. It stinks more of Adobe than YouTube though. They’re not angling to sell ads, they’re trying to appear trendy and relevant.

I tried uninstalling Logi Options+, and then installing SteerMouse, like Stephen did, but SteerMouse doesn’t have the gesture support I am accustomed to. I heard some people used the “Chords” to switch spaces, but I didn’t want to relearn how I used the mouse.

I tried to use Karabiner Elements next. Someone with a more sophisticated background in computer programming might be able to figure that out, but there’s nothing more I could seem to do than what SteerMouse did with setting the thumb button to trigger one thing, and no gestures. If anyone happens to figure out how to reproduce the gestures in Karabiner, get in touch.

Fortunately, Stephen updated his post with some reader feedback that it’s possible to edit the JSON file in Logi Options+ and it won’t run the extra process, or create the tmp directory.

I reinstalled Logi Options+, set up my mouse again (because I have always refused to create a Logi account to sync settings. Like they won’t abuse that), edited the JSON file from true to false and turned off automatic updates.

I had also considered downloading the offline version of the Logi Options+ app that Stephen linked to, but at least with the way things are my mods are very undoable if I do need to update the app.

Like I said in my previous posts about why this crap happens, the people at Logitech talked themselves into how this was actually a good thing that they were doing. Why wouldn’t people want this (poorly implemented) additional feature?

To anyone suggesting that I throw the best mouse I’ve ever used in the trash over this, think again. To people that think I should retrain myself to use a $129 Magic Trackpad set up to the left of my keyboard to switch spaces, I ask, “in this economy?”

This is a case where I want my consumer electronics to be an appliance, not a platform. I’ve gotten it back in line, and that’s that.

Sometimes a mouse is just a mouse.

2024-04-25 11:00:00

Category: text