Federico Viticci

9587 posts on MacStories since April 2009

Federico is the founder and Editor-in-Chief of MacStories, where he writes about Apple with a focus on apps, developers, iPad, and iOS productivity. He founded MacStories in April 2009 and has been writing about Apple since. Federico is also the co-host of AppStories, a weekly podcast exploring the world of apps, Unwind, a fun exploration of media and more, and NPC: Next Portable Console, a show about portable gaming and the handheld revolution.

This Week's Sponsor:

DEVONTHINK

Store, Organize, and Work the Smart Way


On the Future of Vision Pro Inside Apple’s Retail Stores

Earlier today in my Vision Pro story, I wondered about how Apple will showcase and set up the headset for customers in retail stores in the future.

For some excellent analysis on this topic, look no further than Michael Steeber’s latest issue of the Tabletops newsletter. Michael (who’s the leading expert on Apple retail stores) put together some fascinating thoughts on how Vision Pro could marketed and demoed inside the stores, as well as how the product compares to AirPods Pro and Apple Watch from a retail perspective.

Ultimately, the onus of ushering in the era of spatial computing will be on the Specialists and Creatives. The Vision Pro retail experience must be guided from end to end. Apple Stores started as a place to educate, and as technology faded to the background, customers began to intuitively understand their tools and seek out the Apple Store as a product destination. But visionOS is a fundamentally new paradigm that thrusts the role of education front and center once again.

These are just some of the many new challenges and opportunities Vision Pro will bring to Apple Stores. The dawn of spatial computing transforms far more than just the way we interact with software. This new category of device will impel Apple to reshape the retail experience around a more immersive, personalized environment. It’s an incredibly exciting moment.

Check out the concepts and details Michael posted here.

Permalink

Apple Vision Pro: A Watershed Moment for Personal Computing

Vision Pro.

Vision Pro.

I’m going to be direct with this story. My 30-minute demo with Vision Pro last week was the most mind-blowing moment of my 14-year career covering Apple and technology. I left the demo speechless, and it took me a few days to articulate how it felt. How I felt.

It’s not just that I was impressed by it, because obviously I was. It’s that, quite simply, I was part of the future for 30 minutes – I was in it – and then I had to take it off. And once you get a taste of the future, going back to the present feels…incomplete.

I spent 30 minutes on the verge of the future. I have a few moments I want to relive.

Read more


With iPadOS 17, Stage Manager Is (Finally) Moving in the Right Direction

Stage Manager on iPadOS 17 beta 1.

Stage Manager on iPadOS 17 beta 1.

I’m in Cupertino for WWDC this week, and after yesterday’s whirlwind of announcements and surprises, I had some time to sit down with my 12.9” iPad Pro, install iPadOS 17 beta 1 on it, and try the improved version of Stage Manager. As you know, I have a…complicated history with the iPad’s latest multitasking system. Before coming here, I was worried Stage Manager would be left untouched without any updates for at least another year.

I’ll cut to the chase: Apple listened to feedback about Stage Manager and – at least so far – implemented the key improvements I wanted to see. I’ve been using Stage Manager on my iPad Pro since yesterday afternoon, and I even tested it on a portable external display that I brought with me for this trip. If this early, limited experience is of any indication, I think I’m going to be happy with Apple’s revised version of Stage Manager for iPad by the end of the summer. But then again, caution is necessary given how last year’s beta evolved over time.

Read more


Fiery Feeds for iOS Added an In-App Split View Mode That I Wish More iPhone Apps Offered

Vertical split view in Fiery Feeds.

Vertical split view in Fiery Feeds.

A few weeks ago on Mastodon, I shared a simple feature request: a split-screen mode for iPhone RSS readers that would allow me to scroll headlines in the one half of the screen and preview actual articles in the other.

If this sounds familiar, you’re not alone: back in 2007, Steve Jobs demoed pretty much the same thing for the first version of the Mail app for iPhone OS 1.0. That layout mode never shipped, and probably rightfully so at the time given the limited screen real estate of the first iPhone.

Read more


S-GPT 1.0.2 Brings Date and Time Awareness, Integration with macOS Services Menu, Passthrough Mode, Better HomePod Support, and More

S-GPT 1.0.2.

S-GPT 1.0.2.

I just published version 1.0.2 of S-GPT, the shortcut I released last week to have conversations with OpenAI’s ChatGPT and integrate it directly with native features of Apple’s OSes. You can find the updated download link at the end of this post, in the original article, and in the MacStories Shortcuts Archive; before you replace version 1.0.1 of S-GPT, save your existing OpenAI API key somewhere as you’ll have to paste it again in the shortcut later.

I’m going to include the full changelog for S-GPT 1.0.2 below, but long story short: S-GPT is now aware of the current date and time, and I’ve heard all the requests about improving interactions with the HomePod and Siri, so I made that part much better. S-GPT can now perform a variety of date/time calculations with natural language, and you can end a conversation by saying “no” or “stop”.

Read more


S-GPT 1.0.1

I just released a small bug fix update for S-GPT, my shortcut to integrate OpenAI’s ChatGPT large language model with the Shortcuts app on all Apple platforms.

Version 1.0.1 of S-GPT is a relatively minor update that comes with an initial round of improvements based on early feedback I’ve received for the shortcut, including:

  • A proper error-checking alert that tells what went wrong with a request to the ChatGPT API;
  • A better summarization of Safari webpages passed from the share sheet that no longer “hallucinates” results but actually summarizes text extracted via Safari’s Reader technology from any web article;
  • A new behavior for text input on watchOS, which now defaults to dictation rather than keyboard input. I’ve covered this more in detail in today’s issue of MacStories Weekly for Club members.

Additionally, I also realized that the usage tips that S-GPT was displaying every time it asked you to enter some text may have been nice the first three times you used the shortcut, but became annoying very quickly. That was especially true when using S-GPT with Siri in a voice context since they would be read aloud every time. For these reasons, I removed tips and simplified the shortcut’s questions to “What do you want to ask?” and “Want to follow up?”.

In case you missed my introduction of S-GPT earlier this week, you can read the original story here and find out more about how the shortcut works and what it does. I updated the links to the S-GPT and S-GPT Encoder shortcuts in the story to the latest version; you can also find the updated shortcuts in MacStories Shortcuts Archive.

I just released a small bug fix update for S-GPT, my shortcut to integrate OpenAI’s ChatGPT large language model with the Shortcuts app on all Apple platforms.

Version 1.0.1 of S-GPT is a relatively minor update that comes with an initial round of improvements based on early feedback I’ve received for the shortcut, including:

  • A proper error-checking alert that tells what went wrong with a request to the ChatGPT API;
  • A better summarization of Safari webpages passed from the share sheet that no longer “hallucinates” results but actually summarizes text extracted via Safari’s Reader technology from any web article;
  • A new behavior for text input on watchOS, which now defaults to dictation rather than keyboard input. I’ve covered this more in detail in today’s issue of MacStories Weekly for Club members.

Additionally, I also realized that the usage tips that S-GPT was displaying every time it asked you to enter some text may have been nice the first three times you used the shortcut, but became annoying very quickly. That was especially true when using S-GPT with Siri in a voice context since they would be read aloud every time. For these reasons, I removed tips and simplified the shortcut’s questions to “What do you want to ask?” and “Want to follow up?”.

In case you missed my introduction of S-GPT earlier this week, you can read the original story here and find out more about how the shortcut works and what it does. I updated the links to the S-GPT and S-GPT Encoder shortcuts in the story to the latest version; you can also find the updated shortcuts in MacStories Shortcuts Archive.

S-GPT

S-GPT is a shortcut to have conversations with OpenAI’s ChatGPT assistant on your iPhone, iPad, and Mac. The shortcut supports both text conversations as well as voice interactions when used inside Siri. S-GPT comes with native system integrations on Apple platforms including the ability to process text from your clipboard, summarize text found in photos, export conversations to Files and Finder, and even create playlists in the Music app. The shortcut requires an OpenAI API token and a helper shortcut called S-GPT Encoder that needs to be downloaded separately.

Get the shortcut here.

S-GPT Encoder

This is a helper shortcut for S-GPT that needs to be downloaded and installed separately. Without this shortcut, S-GPT won’t work.

Get the shortcut here.

Permalink

Introducing S-GPT, A Shortcut to Connect OpenAI’s ChatGPT with Native Features of your iPhone, iPad, and Mac

S-GPT for Shortcuts.

S-GPT for Shortcuts.

It’s the inaugural week of the second annual edition of Automation April, and to celebrate the occasion, I’ve been working on something special: today, I’m introducing S-GPT, an advanced conversational shortcut for ChatGPT that bridges OpenAI’s assistant to native system features of iOS, iPadOS, macOS, and watchOS.

S-GPT (which stands for Shortcuts-GPT) is free to use for everyone, but it requires an OpenAI account with an associated pay-as-you-go billing plan since it takes advantage of OpenAI’s developer API, which has a cost. S-GPT was built with the latest ChatGPT API, and it can be used both with the existing ChatGPT 3.5 model or – if you have access to it – the ChatGPT 4 API.

While the shortcut is free for MacStories readers, I will be publishing a detailed, in-depth Automation Academy class soon for Club MacStories Plus or Premier members to explain the techniques and strategies I used to build this shortcut. I genuinely think that S-GPT is, from a technical perspective, my best and most advanced work to date; I hope my Academy class will help others learn some useful tips for Shortcuts and, in return, make even better automations for our contest.

With that said, let’s look at what S-GPT is and what you can do with it.

Read more


Using ‘Phone Link’ Between iOS and Windows for iPhone Notifications and iMessage

I enjoyed this explanation by The Verge’s Tom Warren on how Microsoft’s Phone Link app – which has long allowed Android users to connect their smartphones to a Windows PC – has been updated to support iOS notifications and sending texts via iMessage. From the story:

The setup process between iPhone and PC is simple. Phone Link prompts you to scan a QR code from your iPhone to link it to Windows, which automatically opens a lightweight App Clip version of Phone Link on iOS to complete the Bluetooth pairing. Once paired, you have to take some important steps to enable contact sharing over Bluetooth, enable “show notifications,” and allow system notifications to be shared to your PC over Bluetooth. These settings are all available in the Bluetooth options for the device you paired to your iPhone.

And:

Microsoft’s Phone Link works by sending messages over Bluetooth to contacts. Apple’s iOS then intercepts these messages and forces them to be sent over iMessage, much like how it will always automatically detect when you’re sending a message to an iPhone and immediately switch it to blue bubbles and not the green ones sent via regular SMS. Phone Link intercepts the messages you receive through Bluetooth notifications and then shows these in the client on Windows.

I got access to the updated version of Phone Link on my PC today, and this integration is pretty wild and it actually works, albeit with several limitations.

First, the setup process is entirely based on an App Clip by Microsoft, which is the first time I’ve seen and used an App Clip in real life. Essentially, my understanding is that this works similarly to how an iPhone can pair with an old-school Bluetooth car system: the iPhone and PC pair via Bluetooth, and you can then provide the PC with access to your notifications and contacts from iOS’ Bluetooth settings. This is the same UI I have for my KIA Sportage’s system, which uses regular Bluetooth to pair with my iPhone and can also display contacts and missed calls.

The setup process based on an App Clip.

The setup process based on an App Clip.

The difference between my car and Phone Link, of course, is that with Phone Link you can type text messages from a PC and they will be sent as iMessages on iOS. This bit of dark magic comes with a lot of trade-offs (check out Warren’s full story for the details on this), but it works for individual contacts. I’ve been able to start a conversation with John, reply to his messages from Windows notifications, and even send him URLs1, and they were all correctly “intercepted” by iOS and sent over as iMessages. I’ve also been impressed by the ability to clear notifications from a PC and have them go away on iOS’ Lock Screen immediately.

The Phone Link app paired with my iPhone.

The Phone Link app paired with my iPhone.

This was then sent as an iMessage.

This was then sent as an iMessage.

The limitations of Phone Link for iPhone users mean you’ll always have to fall back to the actual iOS device for something – whether it’s posting in an iMessage group or sending a photo or acting on notifications – but for quick messages, glancing at notifications, and clearing them, I think this integration is more than good enough.


  1. Fun fact: raw URLs sent from Windows are delivered as rich links from iMessage, but the card’s preview doesn’t load by default on the recipient’s device. ↩︎
Permalink

Apple Releases iOS and iPadOS 16.4 with New Emoji, Notifications for Web Apps on the Home Screen, Voice Isolation for Cellular Calls, New Shortcuts Actions, and More

iOS 16.4 brings new emoji, push notifications for web apps on the Home Screen , Mastodon link previews, and more.

iOS 16.4 brings new emoji, push notifications for web apps on the Home Screen , Mastodon link previews, and more.

Today, Apple is releasing iOS and iPadOS 16.4, the fourth major updates to the OSes that introduced support for the customizable Lock Screen and Stage Manager last year, respectively.

Ahead of the debut of Apple Music Classical tomorrow and just a few months before a WWDC that’s rumored to be focused on the company’s upcoming headset and a relatively small iOS 17 update, 16.4 is comprised of two big additions to iOS and iPadOS (new emoji and push notifications for web apps on the Home Screen) alongside a variety of smaller, but notable improvements such as some new Shortcuts actions, Mastodon link previews in iMessage, some tweaks to Podcasts and Music, and more.

Let’s take a look.

Read more