OpenAI Unveils Codex “Superapp” Update with Computer Use, Automations, Built-In Browser, and More

Source: OpenAI.

Source: OpenAI.

Today, OpenAI introduced a long list of productivity and coding updates to Codex. I haven’t had a chance to try the new features myself yet, but the demo OpenAI gave me was as impressive as the company’s message was clear: Codex isn’t just for coders anymore.

It was just over a week ago that OpenAI raised $122 billion in financing and announced it was shifting its focus to building a superapp that brings the capabilities of its models into a unified experience. It turns out that app is Codex, OpenAI’s app that, until today, was focused primarily on developing software.

However, according to OpenAI, 50% of Codex’s users were already giving it non-coding tasks to complete. Combined with the OS flexibility of a desktop environment, that made Codex the natural place to bring together a wide range of new productivity and coding features.

On the productivity side of things, the update allows Codex to operate your desktop apps, interacting with interface elements and inputting text, for example. We’ve seen computer use from other AI companies before, but one thing that sets Codex apart is its ability to work in your apps in the background so they don’t steal the focus from whatever app you’re already using.

Codex's built-in browser. Source: OpenAI

Codex’s built-in browser. Source: OpenAI

OpenAI has drawn aspects of its Atlas browser into Codex, too. This allows Codex to prototype websites and apps that users can comment on in-line, creating a tight feedback loop for refining designs. Currently, this feature is limited to running sites and apps via a local server setup, but OpenAI says it will be extended to incorporate actions like interacting with the greater Internet, taking screenshots, and stepping through user flows in the future.

Plugins are taking a big leap forward as well, with over 100 being added to the mix. Like the Claude plugins that Anthropic offers, Codex plugins are composed of a bundle of skills, app integrations, and MCP servers. According to OpenAI, the list includes many popular third-party tools and services like the Microsoft suite, Atlassian Rovo, CodeRabbit, Render, and Superpowers. One of my favorite moments in the Codex demo I saw was a prompt that simply asked, “Can you check Slack, Gmail, Google Calendar, and Notion and tell me what needs my attention?” It’s the sort of query that I think a lot of people can relate to as they start a busy day, and it’s all driven by stacking multiple plugins.

Plugins in action. Source: OpenAI.

Plugins in action. Source: OpenAI.

OpenAI is also testing an enhancement of Codex’s memory feature as a preview that learns from you as you work. Codex will pick up on your preferences, corrections you make, and context from the tasks you give it. This is the sort of feature that is hard to demo, so I don’t have a good sense for it yet, but I expect that over time, its practical utility will become more clear.

One place OpenAI says Codex’s enhanced memory system will help is with new proactive suggestions. As the app learns your preferences and work patterns, it will offer suggestions on what to do next or where to pick up where you left off. Again, how well this will work in practice remains to be seen, but this is exactly the sort of thing that has made OpenClaw so popular. Having an agent that understands your preferences and accesses your messages, files, and other data in a proactive way can be incredibly useful if done well.

Automations. Source: OpenAI.

Automations. Source: OpenAI.

Automations have been expanded, too, allowing Codex to use past threads and schedule tasks over days or weeks. These heartbeat automations stay in the same Codex thread and can be modified by the model itself, allowing it to schedule its own follow-ups – again, very much like OpenClaw.

Also new to Codex is support for gpt-image-1.5 for creating image assets as part of workflows like creating presentations, website mockups, and product concepts.

Developers get new sidebar tools and more. Source: OpenAI

Developers get new sidebar tools and more. Source: OpenAI

Although the focus of today’s update is on productivity, developers haven’t been forgotten. New development features include:

  • Fast frontend iteration using a combination of the in-app browser, computer use, and image generation tools;
  • Multiple terminal tabs;
  • A file sidebar for previewing PDFs, spreadsheets, slides, and other formats;
  • GitHub PR review support, allowing for review of comments inside Codex;
  • A summary pane that tracks plans, sources, and artifacts in a single view; and
  • Remote devbox SSH, an alpha feature for connecting to remote development environments.

That’s a lot, but with more than three million users per week, Codex has proven its popularity well beyond its core coding audience. I’m still skeptical about how much functionality a single app can support, especially when OpenAI addresses the mobile market. I also wonder whether Codex’s productivity and developer tools can coexist without alienating some segment of the app’s users. However, proactive automation of busy work and sifting through mountains of messages and other data is precisely what I’ve wanted from Codex from the start. I’ve seen what it can do when I’m working on a script or app and can’t wait to apply that to my everyday work, too.

Today’s Codex update is available in the desktop app to users with a signed-in ChatGPT account. Computer use is a Mac-only feature at launch (undoubtedly thanks to macOS’s deep accessibility support that was the basis of the same sort of computer use magic we saw in Sky, which was acquired by OpenAI last year), and a rollout of the new features will happen in the EU later. Personalization features like proactive suggestions and the memory enhancements will be coming to Enterprise, Edu, and EU users soon, too.


Google Releases Gemini for Mac

Google released a native Mac app for its Gemini chatbot today.

The app, which can be launched from your Applications folder, Dock, the menu bar, or a global hotkey, will be familiar to anyone who has used Gemini in a browser. The chatbot supports Gemini 3 in Fast and Thinking modes, as well as Pro mode, which uses Gemini 3.1 Pro. Gemini can also interact with files, the contents of a window, Google Drive, Photos, and NotebookLM. It’s multimodal, too, with support for the generation of text, images, video, and music. Dig a little deeper into Gemini’s menus and you’ll find support for Canvas, Deep Research, Guided Learning, and Personalized Intelligence.

A Gemini mini window is available from the menu bar and a global hotkey.

A Gemini mini window is available from the menu bar and a global hotkey.

Even though I just downloaded the app a short time ago, my Gemini chat history was immediately available in the app. The history appears in the app’s sidebar along with a search field, My Stuff, which includes things like images and video generated in the past, and access to your account. The app is written in Swift which was a pleasant surprise.

All my past prompts were immediately available in the new Gemini Mac app.

All my past prompts were immediately available in the new Gemini Mac app.

I’ve only just begun testing Gemini for Mac, but I can already tell that it’s a cut above my hand-crafted single-purpose Safari web app solution. All the same tools found on the web are here, but in a native wrapper, which I appreciate. If you use a Mac and Gemini, the new app is well worth giving a try.

Gemini for Mac is available as a free download from Google.


Introducing Apple Frames 4: A Revamped Shortcut, Support for Frame Colors, Proportional Scaling, and the Apple Frames CLI for Developers

Apple Frames 4.

Apple Frames 4.

Well, it’s been a minute.

Today, I’m very happy to introduce Apple Frames 4, a major update to my shortcut for framing screenshots taken on Apple devices with official Apple product bezels. Apple Frames 4 is a complete rethinking of the shortcut that is noticeably faster, updated to support all the latest Apple devices, and designed to support even more personalization options. For the first time ever, Apple Frames supports multiple colors for each device, allowing you to mix and match different colored bezels for each framed screenshot; it also supports proportional scaling when merging screenshots from different Apple devices.

But that’s not all. In addition to an updated shortcut, I’m also releasing the Apple Frames CLI, an open source command-line utility that lets developers and tinkerers automate the process of framing screenshots directly from the Mac’s Terminal. And there’s more: the Apple Frames CLI is also designed to work with AI agents, and it comes with a Claude Code/Codex skill that lets coding agents take care of framing dozens or even hundreds of screenshots in just a few seconds, from any folder on your Mac.

Apple Frames 4 is the result of an idea I had months ago that enabled me to remove more than 500 actions from the shortcut, going from over 800 steps down to ~300. I did all that work manually, but it was worth it; the improved shortcut is faster and vastly more reliable than before thanks to a more intelligent logic that adapts to the growing ecosystem of Apple screen sizes and display resolutions.

Apple Frames 4 and the Apple Frames CLI represent a substantial step forward for screenshot automation, and I’ve been using both extensively for the past few weeks.

Let’s dive in.

Read more


Podcast Rewind: Text Editors, the Pocket Taco, Photo Backups, an Apple Frames Preview, and an Interview with Evan Ratliff

Enjoy the latest episodes from MacStories’ family of podcasts:

AppStories

This week, John and Federico revisit the state of text editors on Apple platforms and how they use them.

On AppStories+, John and Federico dig deeper into their writing workflows and the apps they’re using to write.

NPC: Next Portable Console

This week, Brendon reports back on the hardware he saw at PAX East, John gives the Pocket Taco a try, and with Federico, they cover the latest handheld news.

On NPC XL, Federico shows off the Lenovo Legion Tab Gen 3, and Brendon shares the SuperStation One.

First, Last, Everything

Jonathan is joined by Evan Ratliff, an award-winning investigative journalist, bestselling author, and podcast host, known for his reporting on technology, crime, and online identity.

Comfort Zone

Chris has some new headphones, Matt wants to talk about why development is still complex in the AI age, and everyone tries to find the most clever way to back up their photos.

On Cozy Zone, the game roasts each other’s ten-year-old Home Screens.

MacStories Unwind

This week, Federico has a sneak peak at a big update to his Apple Frames shortcut, and both he and John share a couple of their favorite Apple TV shows.

Read more



Hour by Hour: Reverse Engineering Your Schedule

Hour by Hour is a clever new approach to scheduling your time from Joe Humfrey of Selkie Design that took me a little while to get used to, but has really grown on me.

The app was inspired by travel planning and the age-old question, “When should I leave for the airport?” You’ve probably been there before. You have a flight at, say, 2:00 pm, but you need to drive 30 minutes to the airport, add some time to park, take a shuttle to the terminal, get through security, and build in a little extra wiggle room just in case traffic is bad or something else goes sideways. Suddenly, 2:00 pm becomes an exercise in mental gymnastics as you work your way back to when you should walk out the door.

Hour by Hour solves this sort of scheduling, but for every type of event, by using the same kind of reverse planning. At the same time, it’s not really a calendar app so much as a scheduling companion for your calendar. You can pull your calendar events into Hour by Hour, but you don’t have to, and if you dive into the app expecting to use it the same way you use a traditional calendar, the assumptions you bring with you will probably trip you up.

Read more


Steam Announces Steam Link for Apple Vision Pro and Releases Beta Version

The gaming ecosystem on visionOS continues to grow, this time with the announcement that Steam Link is coming to the platform. The official app will allow users to wirelessly stream games in 2D from Steam on their local Macs and PCs to their Vision Pros. According to the company, the app can stream games at up to 4K resolution and includes an adjustable, curved panoramic view. A TestFlight beta is now available for those who want to test the app before its public release.

Steam’s stature in the PC gaming market cannot be understated, so opening the platform up to the Vision Pro is a huge boon for gaming on the device. Of course, one would hope to see VR streaming support come with time, but official support even for 2D gaming is a big step.

Between an official Steam client, 4K cloud game streaming from NVIDIA GeForce NOW, console remote play via Portal, local OpenVR game streaming with ALVR, HDMI input over UVC via the Developer Strap, and native titles from Apple Arcade and the App Store, the variety of games available for Vision Pro users to play is getting larger by the day. There’s lots to be excited about in the world of visionOS gaming these days.


LunarWall: Shuffle Moon Photos from Artemis II On Your Lock Screen or Mac Desktop

LunarWall for iOS.

LunarWall for iOS.

I’ve been staring at my Lock Screen and macOS desktop a lot this week. Not because of John’s iMessage notifications or the weird handhelds we share in the NPC group thread – because of the Moon. Specifically, because of photos taken by Orion as it swung within 4,067 miles of the lunar surface during the Artemis II flyby a couple of days ago. Yesterday, NASA published an official gallery of images from the flyby, and I immediately knew what I had to do.

LunarWall is a simple shortcut that picks a random image from a curated set of 23 photos pulled from NASA’s Artemis II Lunar Flyby gallery and sets it as your wallpaper. That’s it! Each time you run it, you get a different photo. The way this shortcut works, NASA’s images aren’t re-hosted or saved anywhere on your computer: the LunarWall shortcut fetches each image directly from NASA’s CDN and passes it to the ‘Set Wallpaper’ action, which is configured to automatically crop images to fit on mobile devices, blurs the wallpaper for the iOS/iPadOS Home Screen, and uses the original widescreen images at high resolutions on macOS.

Read more


Claude Mythos Preview Will Only Secure Part of the Internet

Yesterday, Anthropic announced Claude Mythos Preview, a new general-purpose model that it says is exceptionally good at finding security vulnerabilities in code. In fact, the model is so good that Anthropic has decided not to release Mythos Preview to the general public. Instead, it’s being released to a select group of companies that control OSes and other critical software.

Anthropic found thousands of vulnerabilities across every major OS and web browser with Mythos Preview, but used these three examples to illustrate their severity:

  • Mythos Preview found a 27-year-old vulnerability in OpenBSD—which has a reputation as one of the most security-hardened operating systems in the world and is used to run firewalls and other critical infrastructure. The vulnerability allowed an attacker to remotely crash any machine running the operating system just by connecting to it;
  • It also discovered a 16-year-old vulnerability in FFmpeg—which is used by innumerable pieces of software to encode and decode video—in a line of code that automated testing tools had hit five million times without ever catching the problem;
  • The model autonomously found and chained together several vulnerabilities in the Linux kernel—the software that runs most of the world’s servers—to allow an attacker to escalate from ordinary user access to complete control of the machine.

A lengthy Frontier Red Team report brings the receipts for security researchers with an in-depth look at what Mythos Preview uncovered and the step change that the new model represents over Opus 4.6:

For example, Opus 4.6 turned the vulnerabilities it had found in Mozilla’s Firefox 147 JavaScript engine—all patched in Firefox 148—into JavaScript shell exploits only two times out of several hundred attempts. We re-ran this experiment as a benchmark for Mythos Preview, which developed working exploits 181 times, and achieved register control on 29 more.

As part of a test, Mythos Preview also managed to escape its sandboxed environment, message the researcher conducing the test, and then, outside the parameters of the test, posted about the exploit online.

The idea behind Project Glasswing, whose participants include Amazon Web Services, Anthropic, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, the Linux Foundation, Microsoft, NVIDIA, and Palo Alto Networks, is to give them a head start at securing their systems before similar models emerge and are exploited for cyberattacks. If Mythos Preview’s capabilities are as Anthropic makes them out to be, this seems like the right approach. However, I do worry that with time, it could lead to a two-tier Internet where big tech companies operate in relative security thanks to tools like Mythos Preview, while those without access are left to swim with the sharks.