The AI Daily Brief: Artificial Intelligence News and Analysis
The AI Daily Brief: Artificial Intelligence News and Analysis

How to Use Claude's Massive New Upgrades

2h ago25:365,255 words
0:000:00

Claude’s latest updates aren’t just incremental—they fundamentally change how you interact with AI, shifting from tool to always-on execution layer. This episode breaks down the biggest new capabiliti...

Transcript

EN

Today on the AI Daily Brief, how to use all of Claude's massive new upgrades.

The AI Daily Brief is a daily podcast and video about the most important news and discussions in AI.

Alright, friends, quick announcements before we dive in.

First of all, thank you to today's sponsors, recall.au, assembly, prompt QL and Blitzie. To get an ad free version of the show go to patreon.com/aideallybrief or you can subscribe it up a podcast to learn about sponsoring the show. Send us a note at [email protected]. We are getting very close to full with Q2 spots.

So if you're promoting for example a launch or anything in the near future, it is definitely a good time to reach out. Now, one more note on today's show. Yesterday it was kind of a big politics and society episode. Today we're running a whole way in the other direction with something a lot more practical.

There have been so many new things that have come out for Claude, Claude, Claude co-work. That was time to go through them all and ended up running quite long so we will end up not doing the headlines today. We will be back with a normal headlines episode tomorrow. Two more things to quickly flag before we get into that.

First agent madness is live, the round of 64 is going, hundreds of you have voted

and voting will close for this first round at the end of the day on Thursday, March 26th.

So go to agent madness.au to check that out. And finally, I'm noticing that I'm doing a lot more little companion experiences with these episodes and so as of today, we're launching play.au daily brief.au which is where all of those fun little experiences live. So for example, the day you can find the checklist of things to try with Claude right now

and again, that'll live at play.au daily brief.au. With all that out of the way, let's talk about all the cool new things you can do with Claude. Every other day or so for the last month, there has been some part of the headlines where Claude launched some new feature or upgrade for Claude code or co-worker some other part of the Claude ecosystem.

And at this point, it is officially gotten to the point where we needed to do a full-on retrospective of everything that has launched over the last month or so to help you guys map out and figure out how to use all of these big new upgrades. Now, of course, this comes in a specific lineage of updates. Step one was the models.

The Opus 4.5 GPT-5 2 generation of models that came about at the end of last year, catapulted us into a new capability era that Opus 4.6 GPT-5 3 codecs and GPT-5 4 have continued. Step two, of course, was the unlock that Opus Claude represented. It was a harness that brought with it a whole bunch of concepts and user interaction patterns and behavior sets that made building your own agents and agent teams all of a sudden more viable

and more realistic. Step three has been the absolute race ever since Opus Claude blew up to bring those types of features to all the other AI products. This qualification trend has been one of the big themes ever since Opus Claude launched. Now, when Opus Claude's founder Peter Steinberger was hired by OpenAI,

some people jumped in to say that Anthropoc had made a big goof by not bringing him over there. Others' response was a little bit closer to, "I don't know, man. Let's let him cook and see what happens."

And certainly you have to think that the people saying let them cook are feeling pretty finicated right now.

The qualification of Claude, maybe we'll call it the Claudeification, kicked off at the end of February with remote control. Remote control was a way to bring Claude codes specifically to your mobile experience. Both the mobile capabilities of Opus Claude, as well as its ability to bridge between different types of devices, were some of the parts of that system that people were most excited about.

And so it wasn't all that surprising to see Claude code jump on that first.

What's more, even outside of Opus Claude, this is such a natural extension of the product that you got to imagine that this would have happened anyway. The way that remote control works is that you start your task in your Claude code terminal session and then you can pick it up and continue working from your phone. Now there's nothing happening in the cloud here.

Basically, when you've started that remote control session on your machine, Claude is going to continue to run locally that entire time. That gives remote control the ability to use your full environment, including your file system, MCP servers, tools, and product configurations, and you can go back and forth interchangeably.

In the docs they write, unlike Claude code on the web, which runs on cloud infrastructure, remote control sessions run directly on your machine and interact with your local file system. The web and mobile interfaces are just a window into that local session. There are three ways to start a remote control session. You can start a dedicated remote control server by navigating to the specific

product directory you want to work on and running the Claude remote control command from their Claude will display a session URL that you can use to connect from another device. You can also press your spacebar to get a QR code that you can access from your phone. An interactive session basically means that you have your option of going back between using terminal as you normally would with Claude code or using the remote session.

So basically the difference with server mode is that on server mode, you're just using your mobile device to control Claude code, whereas interactive you can go back and forth. Finally, if you are already in a Claude code session and you want to move to remote, you can use the slash remote control command or slash RC,

and this is going to once again display either a URL or QR code that you can use to connect from another device. First impressions were positive. Prominent solo per newer Peter level's road, Claude remote control is extremely nice.

Can edit on Mac OS or iOS and Claude app on my production server from anywhere.

He basically compared it favorably to an SSH session,

which would be another more technically complex way to log into your local device to control it while on the go. Roman Mirzoyan writes, "Yesterday I fixed two bugs and then released an app update to the app store without touching my laptop and having a walk for half a day. Now, as time has gone on, people started to realize that this was a bigger shift that they might have originally thought."

Gagganceludia writes, "Claude code remote control just clicked for me. You kick off a task in the terminal, then pick it up from your phone on a walk. That's not a productivity feature. That's a relationship shift. You stop thinking of it as a tool you operate and start thinking of it as something you delegate to and check in with. Different mental model entirely.

I think that that's right and I think that most people are still just slowly coming to that

realization because it's one you kind of have to live not just hear about."

Next up, a couple weeks later we got a different way to interact with Claude from afar. This feature was for Claude co-work and called dispatch. Anthropics Felix Resparer writes, "We're shipping a new feature and Claude co-work as a research preview that I'm excited about." Dispatch. One persistent conversation with Claude that runs on your computer, message it from your phone, come back to finish work.

Felix then went on to explain a little bit further. Because it's co-work he writes, Claude runs code in a sandbox on your machine. Your files stay local, you approve what Claude touches before it acts. It feels pretty magical to give Claude a mission on my computer and get occasional updates like creating reports from internal dashboards or finding me a better seat on my next flight. Everything Claude can do on your computer, files,

browsers, tools are reachable from wherever you are. Now one constraint Felix writes is that your desktop has to be running. Claude co-PM know as we've been also talked about dispatch.

Coolest abilities he writes? One send files from local machines so you can work on

PowerPoints on the go, two spawn subsessions on desktop that you can drill down on, three chat about any local co-work session. In the docs they explain a little bit more about how this works. In topic writes, instead of starting a new session for each task you have a single persistent thread with Claude. This thread doesn't reset. Claude retains context from previous tasks so you can pick up where you left off. Message Claude from your phone on the way to work,

then follow up from your desktop when you sit down. It's the same conversation, same context wherever you reach it. When you assign a task, Claude figures out what kind of work is needed and spins up the right session. Development tasks run in Claude code, knowledge work runs in co-work. These sessions appear in their respective sidebars. You can click into any session for details or wait for the result in the thread. Claude messages you the outcome, a spreadsheet, a memo,

a comparison table, a pull request, rather than showing you every step of the process. You'll get a push notification on your phone when a task is done or when Claude needs your go ahead. Now, like with remote control, the power users who really started to dig into dispatch

found that it was not just the shift in scale but a shift in kind.

Pavel Huron writes, "Dispatch didn't fill my dead time, it changed how I structured my day. I went to the jump arena with my kid because I could direct work async from the sidelines. The model isn't grind during gaps. It's designed your day differently because the work runs without you sitting in front of it. He also wrote an article after 48 hours of experimenting. And one thing that he points out is that dispatch is not Claude chaught on your phone.

Dispatch he writes is an orchestrator. From a single conversation on your phone, you spawn and manage multiple cowork task sessions running simultaneously on your desktop. Each session runs independently. It's own context, it's own file access, it's own connectors. Your phone is the command chair. Your desktop does the heavy lifting. Think of the difference between texting someone a request and sitting in a control room with multiple

screens. Each screen is a task session running on your desktop. Your phone directs them all from one conversation thread. So what are the types of tasks he did? During morning coffee at home, Pavel started with pull the latest competitor updates and summarized changes since last week, as well as draft the sponsor collaboration page using a notion database. While he was walking the dog, he checked task one, the competitor summary. He followed up with added comparison table

against our current roadmap. The readerek he points out took 10 seconds one handed. While in the passenger seat with his wife driving, he reviewed the new sponsor notion page. To formal he said, pull the engagement metrics from the last campaign and make the value proposition sharper. He also started task three, gap analysis on the article draft. While at the jump arena with his kids, he started working on infographic iterations. Move the icons left, change the color of

the third section. Finally, when back at his desk, everything was waiting and he reviewed, adjusted and shipped. Ultimately, he writes the actual direction time across all of these gaps was maybe 25 minutes total. The clawed execution running in parallel was three plus hours of work. Even malloc also had a positive experience with dispatch. He writes after using it a bit, clawed cowork dispatch covers 90% of what I was trying to use open claw for, but feels

far less likely to upload my entire drive to a malware site. What I like better, he writes, easy, much more stable and safe, existing connectors mean better integration with Gmail, browsers et cetera, very good tool use. What's missing for me? The ability to invite clawed to any channel, the heartbeat and productivity and the multiple sessions.

Why is there always a meeting bot in your zoom call? Blame recall.au. Recall.au powers the

Meeting bots and desktop recording apps behind products like Cluey, HubSpot a...

They handled the hard infrastructure work capturing clean recordings, transcripts, and metadata across Zoom, Google Meet, Microsoft Teams, in-person meetings, and more, so developers don't have to build with themselves. If you're building a meeting note taker or anything involving conversational data, recall.au is the API for meeting recording. Get started today with $100 in free credits at recall.au/aidb. That's recall.au/aidb. If you're building anything with Voice AI,

you need to know about assembly AI. They've built the best speech to text and speech

understanding models in the industry, the quiet infrastructure behind products like granola,

dovetail, ashbe, and clueless. As I've said before, Voice is one of the most important modalities

of AI. It's the most natural human interface, and I think it's a key part of where the next way of innovation is going to happen. Assembly AI's models lead the field in accuracy in qualities so you can actually trust the data your product is built on, and their speech understanding models help you go beyond transcription, uncovering insights, identifying speakers, and surfacing key moments automatically. It's developer first, no contracts pay only for what you use and scales

effortlessly. Go to assembly AI.com/brief, grab $50 in free credits, and start building your Voice AI product today. If you're an operator, your day is a non-stop stream of decisions, and most of them require you to look at the data. You don't need another dashboard. You need answers you can trust, fast, but the bottleneck is always the same. The data isn't ready, it's scattered, it's messy, definitions aren't clear, you're reading on your data team,

or waiting on domain experts for clarification and confirmation. That's the bottleneck today's

sponsor, Promp's QL, is built to break. Promp's QL is a trusted AI analyst for high-frequency decision-making. It connects across warehouses, databases, SaaS, and internal APIs. No massive data-prepresent realization required. It's built for multiplayer input. teammates can jump into a thread, correct assumptions, and nuance, flag edge cases. Promp's QL turns everyday conversations into a shared context, and if something is ambiguous, it doesn't guess. It escalates to the right

expert, captures the correct logic, and gets it right next time. That's how it delivers trust

in accuracy. Over time, Promp's QL specializes to your business, like that veteran employee who just knows things. From simple what is questions to complex what if scenarios, you can model impact and stress test decisions before you commit, all through a simple natural language prompt. Promp's QL, the trusted AI analyst for teams with shared context, and messy data. You've tried in IDE copilates, they're fast, but they only see local silos of your code.

Leverage these tools across a large enterprise code base and they quickly become less effective. The fundamental constraint? Context. Blitzy solves this with infinite code context, understanding your code based down to the line level dependency across millions of lines of code. While copilates help developers write code faster, Blitzy orchestrates thousands of agents that reason across your full code base. Allow Blitzy to do the heavy lifting,

delivering over 80% of every sprint autonomously with rigorously validated code. Blitzy provides a granular list of the remaining work for humans to complete with their copilates.

Tackle feature additions, large-scale refactors, legacy modernization,

greenfield initiatives, all 5x faster. See the Blitzy difference at blitzy.com. That's BLITZY.com. Now, if that was all, that would already be a huge amount of new interactive modalities to ingest and change how we were interacting with cloud co-work and cloud code. But we are not done yet. Next up, we got cloud code channels.

Which is basically cloud codes answer to the open-claw interaction of talking to

it via telegram. And through topics to recrites, we just release cloud code channels which allows you to control your cloud code session through select MCPs, starting with telegram and discord. In the docs they write, a channel is an MCP server that pushes events into your running cloud code session, so cloud can react to things that happen while you're not at the terminal.

Channels can be too way. Cloud reads the event and replies back through the same channel like a chat bridge. Unlike integrations that spawn a fresh cloud session or wait to be pulled, the event arrives in the session you already have open. When Larry Valez asked, "How is this better than dispatch?" Or it's just an alternative to dispatch? To recrites? We want to give you a lot of

different options on how you talk to cloud remotely. Channels is more focused on devs who want something hackable. Now, just to add a little bit of clarity here, we see telegram and we just think about ourselves chatting with cloud code. But the whole idea of channels is that it's not just chats from you, yes that is one type of interaction, but you can also connect to other services like, for example, Sentry Monitoring,

directly to cloud code via those channels in telegram or discord, so that cloud can react even while you're not there. Damien Galarza writes, "Instead of cloud pulling data via tool calls, channels push events into the session from the outside. CI failures, web hook payloads, monitoring alerts, chat messages from telegram or discord. Anything that can send in post can now reach your running session." And it definitely seems like of these, "Channels is on the

farthest end of the technical spectrum." Dario and Twitter writes, "If you missed it, channels are

Essentially MCP servers that push events into a cloud code session, letting c...

outside world beyond the terminal." This was the exact missing piece I needed for an idea I've been brewing. My goal was to fold, build a custom orchestration system to spawn cloud code sessions anywhere, Docker VMs, pods, or locally on my MacBook, and then create a custom app across macOS, iOS, and iPad OS to control these agents on the go. Fast forward to today, just for days after channels dropped, and the iOS app is already up and running. I have so many more ideas on how to

tweak and expand this to fit my exact workflow. Now at this point, you'd be forgiven for getting a little confused because we're talking about all these new ways to use cloud core and cloud code remotely, all with slight variations on the theme. This is kind of why you're going to want

to go hack at all of these to understand how they fit your own use cases, but I think one big

takeaway is an aggregate massive shift in how anthropic is imagining how you're going to be

working with cloud in the future. And effectively, the idea is an always-on context-maintaining

persistent interactive orchestration experience where work is happening all the time even when you're not doing it with the right way to interact with that work from wherever you are and whatever your use case is. In addition to those features, we also saw anthropic go after parity with open claw when it came to scheduling tasks. At the end of February, cowork got scheduled tasks, think a morning briefing, weekly spreadsheet updates, or Friday team presentations,

then about a week and a half later, we got local scheduled tasks in cloud code desktop to reeksend that his favorite use cases to ask it to check error logs every few hours and create PRs for any actionable errors. Now because these tasks are local, they run as long as your computer is awake, but then a couple weeks after that, we got recurring cloud-based tasks. No, as we've been again rights, set a repo, or repose, a schedule, and a prompt.

Cloud runs it via Cloud Infra on your schedules so you don't need to keep cloud code running on your

local machine. He said it so far the cloud code team had found it useful for things like sweeping through open PRs, building features from approved issues, analyzing CI failures overnight, and syncing docs based on newly merged PRs. And if you're wondering if this had to do with open claw, in another tweet, no a said, "Some might say I built this because I couldn't figure out how to set up my Mac mini." And yet, in terms of sheer consumer excitement, all of these were just

prelude to what was announced on Monday night, the official claw to count on Twitter rights. You can now enable Cloud to use your computer to complete tasks. It opens your app, navigate your browser, fills in spreadsheets, anything you do sitting at your desk.

About 16 hours after that tweet went live, 40 million people have viewed it.

62,000 people have booked it, and the chatter is overwhelming. Anthropics Felix Reesberg writes, "Today we're releasing a feature that allows Cloud to control your computer, mouse, keyboard, and screen giving it the ability to use any app." Now, he also points out

that these features shouldn't be viewed in isolation. Felix writes, "I believe this is

especially useful if used with dispatch, which allows you to remotely control Cloud on your computer while you're away." And there are announcement blog and topic rights. When Cloud doesn't have access to the tools it needs, it will point click and navigate what's on your screen to perform the task itself. It can open files use the browser and run dev tools automatically with no setup required. When using your computer, they say, "Clawed will reach for

the most precise tool first," starting with existing connectors to services like Slack or Google

Calendar. However, when Cloud finds that there is not a connector set up, it can control browser mouse keyboard and screen to complete different tasks. Now, as you might be thinking right now, this kind of supercharges those features like dispatch. Anthropic agrees, writing, "With dispatch you can tell Cloud to automatically check your emails every morning or post some metrics every week, or spin up a cloud coerc or cloud code session for a report or a pull request."

Cloud's new computer use capability makes dispatch even more helpful. Now, Cloud can use your computer on your behalf while you're away. For example, to create a morning briefing while you're on the train, make changes in your IDE, run tests, and put up a PR, or keep your 3D printing project moving according to your initial plan. A lot of the initial response was fairly breathless. Sweet Birdabow writes, "Clod can now control your entire computer with one prompt,

forever. You tell it scan my email every morning once and it just does it, forever." It can also open apps at its spreadsheets, move files, batch process 150 photos in Photoshop, export PDFs, all of it. You can also start a task from your phone, go to dinner, come back to finish work like nothing happened. Anthropic is giving us the full desktop agent that uses the actual screen mouse and keyboard. Not a sandbox, not a simulation, your real

Jarvis. This is absolutely insane. The LLM era is over. Gagan again, who we heard from earlier, writes, "Bentesting it in co-work mode. The thing that gets me isn't the automation. It's that it figures out what you actually meant when you said handle this for me. That's the shift." Bialara writes, "Clod is no longer just a tool that uses the computer. It operates inside it as a true execution layer. It can replicate everything a human does. Mouse movements,

keyboard input, screen interaction. It can open any application navigate through it and produce the exact output you want. This goes beyond API's. This is full real-world system control. Which means, from anywhere, you can control your devices and automate workflows and to end.

In terms of examples, a lot of people started fairly simply.

"Okay, this is pretty cool. Just had it grab a photo off my desktop from afar.

Dumb and simple, yes, but done a natural language with no weird setup."

Daniel Sahn writes, "I'd only posting about anything I haven't tried myself, so I saw the announcement that Claude dispatched now controls your computer, and I simply asked it to open Twitter and like a Claude post. Done. This is getting more interesting and more dangerous by the

minute. I teach Sheth writes, "This is the first time a mainstream AI product actually behaves

like a junior ops hire that lives on your laptop." And Peter Gustav points out that this could be especially valuable for corporate stuck-in legacy software that isn't going to port into AI ecosystems natively. He writes, "This is another domino to fall. Computer use of arbitrary apps, not just your browser. This is a big deal for a lot of corporates who have custom crappy apps from 20 or 30 years ago. And summing it all up, box is air and leavey argues that this is

to put it simply a big deal. In a long post on Twitter he writes, "Computer use in the ability to write and run code on the fly are the ultimate primitives for agents to be able to take on more and more tasks and knowledge work. Most work requires hopping between multiple applications and working with broad sets of data in a workflow, and agents will need to be able to reverse these systems to be able to effectively automate any real work in the enterprise. Now we will have

agents that are the equivalent of having an expert programmer or any number of them that can write code or use any API to automate whatever work you're doing. Agents will have access to either a user's computer and resources or their own sandbox to operate in and be able to pull together the tools necessary to perform the task at hand. This opens up the broadest set of agent to use cases. To be sure, there are going to be various hurdles around security, permissions

and access control, identity challenges, and more. For instance, should the agent always act on

behalf of the user or should they have their own identity and limited set of access rights?

How do you triage security events when historically volume of activity on a system is no longer a reliable signal of a security issue? How do you ensure the agent isn't going rogue or getting prompted to do something risky? All problems that need to get figured out. Then there's also lots of work needed to ensure software is set up to enable agents to operate with their tools in a headless fashion. This will be an uncomfortable reality for some incumbents and equally a

welcome one for tools that historically have operated seamlessly via APIs and have business models to support this. Lots of change coming in the world of work agents and it's going to get pretty wild. Now we're already going pretty long, but I want to point out that this is not the end of what has been pushed over the last month. Running quickly through some of the second tier and quality of life type of announcements, one big complaint that lots of folks have had about

co-work is that there wasn't a conception of a project. Now there is, which will make a huge difference for day-to-day functionality. Also in March, the Claude code team launched code review, whereas they put it Claude dispatches a team of agents to hunt for bugs. This past month,

the 1 million token context window actually became available generally for both Claude

Opus and Claude's on it. Why commentator's carry tan wrote? I underestimated how powerful Opus 4.6 with 1 million tokens is. Even last year, we were absolutely hitting context limit problems constantly. 1 million tokens mean you can do much more complex analysis entirely in context. Claude code is so much better. Meanwhile, over in the main Claude app, Claude can now build interactive charts and diagrams, and we also got upgrades for Claude for Excel and Claude for

PowerPoint, including the ability to integrate them together. They write, "When you've got more than one file, Opin Claude shares the full context of your conversation between them." Pull data from spreadsheets, build out tables and updated deck without re-explaining a step. Skills are also now available inside the Excel and PowerPoint add-ins, meaning that when your team has some standard workflow, like running a variance analysis or building a client deck,

you can save that as a skill. Other people in your organization can then run it in one click from their own sidebar. For my free users out there, memory and connectors are now both available on the free plan, and we also got a whole plugin marketplace for enterprise customers. One implication of this is that it's pushing a lot of people to think differently about just how productive a small team can be. Ethan Malik again writes, "The ability of the Claude

team to learn from things like Opin Claude and implement features like this on a daily basis is a very strong argument that for AI-powered coding teams, a very different software development

process is possible with large strategic implications." I think for most of us though,

it's just a big ol' checklist of things we need to try. So to quickly sum up, we've got computer use, which allows Claude to control your computer, doing things even in non-native apps, dispatch which allows you to initiate sessions and orchestrate tasks from Claude co-work from your phone, remote control which allows you to interact with Claude code while on the go, channels which allow you to port events impacting your software automatically into Claude code,

via channels like Telegram and Discord, and Schedule Tasks which allow you to have things happening in co-work or Claude code on a scheduled basis either locally or in the cloud. Now for those of you who want to keep track of this simply, you can go find this checklist with links to information about all these features at play.aidalybrief.ai. So many episodes now have these little companion experiences that I'm going to just be putting them there. Again,

let's play.aidalybrief.ai. For now, however, that is going to do it for this episode of the

A.

Compare and Explore