Hi {{first_name}}!

This was one of the most consequential weeks in AI in a long time. We just saw three major labs ship flagship models in a single week. At the same time, Amazon committed a total of $13 billion to Anthropic. This is the largest bilateral AI infrastructure deal ever signed: a signal that AI has moved from a shiny new tool to a permanent utility like the power grid.

Here is what is inside this week:

  • New and Noteworthy: Critical updates from Anthropic, OpenAI, and Google Workspace.

  • The Gemma 4 Test: A local AI model you should try if you handle sensitive data.

  • Tool Spotlight: Granola, the meeting notes tool I use every day.

  • The Agent Audit: My framework for knowing exactly how much to trust your AI agent.

OK, let's get into it!

New and Noteworthy

  • Google Workspace gets a major upgrade. Google rolled out Workspace Intelligence this week. If your team uses Google, Gemini now has real-time access to everything across your Gmail, Drive, Chat, and Calendar. This removes the need to copy context or explain projects manually. It is turned on by default, so if you are a Workspace admin, check your console to manage which data sources Gemini can search. Your data is not used for model training or advertising.

  • Cowork builds live dashboards. Claude in Cowork can now build "live artifacts." These are dashboards and trackers that stay connected to your apps and files. When you open them, they refresh with current data instead of giving you a static snapshot.

  • OpenAI launches ChatGPT Images 2.0. This is a new "reasoning" image model called gpt-image-2. Unlike older tools, it works through a visual problem before generating the final image. The result is better consistency, 4K resolution output, and near-perfect text rendering. This is a huge win for creating marketing assets or presentations without the usual AI glitches. It’s early but ChatGPT seems to have dethroned Gemini’s Nano Banana as the best image generation model (for now). 

  • Claude Opus 4.7 is live. Anthropic’s most capable model is now available. It is a meaningful step up for complex work like financial analysis, multi-step research, and document creation. It also features a massive context window, meaning it can hold an enormous amount of information in a single session without losing track of details.

  • Claude Design is here. Anthropic launched Claude Design this week, a new canvas-style interface with chat on the sidebar for building wireframes and high-fidelity prototypes. It starts by asking you 5 to 10 questions through an interactive form, then gets to work. I've been testing it and the image-to-design workflow in prototype mode is amazing, you can drop in a reference image and it builds from there. It's currently in research preview with separate limits, so expect to get 2 to 3 full generations per week on the $20 plan before hitting the ceiling. Worth experimenting with before the limits decrease.

  • Salesforce turns your CRM into an AI engine. Salesforce launched Headless 360. This turns your customer data into a live layer that AI agents can query and act on in real time. This is Salesforce positioning itself as the data backbone for your future AI operations.

  • Anthropic and Amazon just signed the largest bilateral AI infrastructure deal on record Amazon injected a fresh $5 billion into Anthropic this week, bringing their total committed investment to $13 billion. In exchange, Anthropic pledged over $100 billion in AWS cloud spending over the next decade, securing 5 gigawatts of compute capacity to train and run Claude. The bigger signal here: this isn't a startup funding round, it's infrastructure at a scale that rivals power grids. For businesses already using Claude through AWS, this means more reliability, faster performance, and expanded capacity in Europe and Asia.

  • Google's Sergey Brin sounded the alarm internally: Claude is winning the coding race A leaked memo from Google co-founder Sergey Brin to the company's AI team warned that Google "must urgently bridge the gap in agentic execution" to compete with Anthropic's Claude Code. Both Brin and Google DeepMind's CTO are now directly involved in a new internal strike team focused on catching up. This matters for non-developers too, because whichever AI platform wins the coding race will likely win the broader enterprise AI race. Right now, Anthropic holds the lead in agentic coding, at least according to Brin's own assessment, and the rest of the industry is reacting to that.

  • OpenAI is expanding its agents SDK: more autonomy, better sandboxing. OpenAI quietly shipped a significant upgrade to its Agents SDK this week, giving developers better infrastructure to build agents that can work across files, run commands, edit code, and complete multi-step tasks autonomously. The upgrade includes native sandbox execution, so agents work inside a controlled environment rather than running loose. This is more of a developer story, but the downstream effect hits every business: the AI-powered tools your vendors are building are about to get even more capable, and significantly more automated.

I've Been Testing Gemma 4

Last week I mentioned Gemma 4, Google's latest open-weights model family, and a handful of you hit reply asking for more. So let's go a little deeper.

Gemma 4 is different from most model releases you'll hear about: it's not built for the cloud. It's built to run on your device, meaning your data never leaves your machine. For business owners handling client information, financial records, or anything sensitive, that's not a small thing.

The family comes in different sizes depending on what you're working with. The smaller versions are designed for laptops and phones. The larger ones run on a single consumer GPU or a standard workstation. You're not buying server infrastructure to use this. You're running it on hardware you likely already have.

Here's what's been interesting to watch in practice.

The model handles text and images natively, and gives you real control over how it processes visual input. If you need speed, you dial down the visual detail. If you need accuracy on something like reading a dense document or parsing a chart, you dial it up. 

Gemma 4 is also designed to call functions and tools, not just generate text. That means it can fit into an agentic workflow where it's actually doing things, pulling data, triggering actions, completing steps, rather than just answering questions.

Gemma 4 is not going to outperform Claude or ChatGPT on complex reasoning tasks. That's not what it's for. But if you're building a workflow that handles sensitive data, runs repetitive processes, or needs to operate offline, a well-configured Gemma 4 setup can punch significantly above its weight.

I’ve installed their smallest model (Gemma e2b) on my iPhone 17 and I'm continuing to test this and will share a more hands-on breakdown once I have more to show. If you're already experimenting with local AI deployments or curious about where to start, hit reply. I want to hear what you're working on.

Tool Spotlight: Granola

A few clients asked me what I use for meeting notes. The answer right now is Granola.

It sits quietly on your computer, no bot joining the call and just transcribes everything in the background. When the meeting ends you get clean, structured notes with action items already pulled out.

What sets it apart from most note-taking tools is the notepad. During the call, you have a live writing space where you can jot down your own thoughts, reactions, and anything you want to flag. When the meeting ends, Granola takes what you wrote and weaves it together with the full transcript into one clean set of notes. So it's not just an AI summary of what was said, it's your perspective combined with everything that happened which is a meaningful difference.

You also stay in control of when it runs. You can choose to have Granola transcribe a call or skip it entirely, which matters when you're jumping into a conversation that doesn't need to be recorded. Not every meeting should be captured and it's good that the tool respects that.

One thing to know before you roll Granola out to your whole team: by default, notes are shareable via link and AI training is opted in. 

Granola is free to start. Business plan is $14 a month. Available on Mac, Windows, and iPhone.

The Agent Audit: Know Exactly How Much to Trust Your AI Before You Let It Run

If you've started experimenting with AI agents, you've probably hit this moment: the agent does something impressive, and you wonder how much you can actually hand off to it. It's a good question, and honestly one most people are figuring out in real time.

Here's a framework I've been using with clients inside Claude Cowork that takes the guesswork out of it.

Every AI agent you deploy starts at 20% trust. That number goes up when the agent executes cleanly and anticipates what you need. It goes down when you have to repeat yourself, catch errors, or redo its work. The key is only giving the agent tasks that match its current trust level:

  • 20-40%: You review every output before anything goes out. Good for anything touching clients, money, or decisions you can't reverse.

  • 40-60%: The agent works independently but checks in at key decision points. You're not watching every move, but you're still in the loop on what matters.

  • 60-80%: The agent runs end-to-end and you spot-check results the next morning. It's earned enough trust to operate, but you're still auditing.

  • 80%+: The agent decides when to ask. Hands off.

The part that makes this work is the nightly Agent Audit. At end of day, set up a Claude Schedule Task in Cowork with this prompt:

Review everything you completed today. For each task: (1) rate your execution quality 1-5, (2) flag any output where I had to correct you or repeat myself, (3) propose one update to your instructions that would prevent that correction next time, and (4) tell me whether your trust level should go up, down, or stay the same. Output a single summary I can review.

Every morning you get a clear picture of whether your agent earned more autonomy overnight and because it's updating its own instructions, it keeps improving without you rewriting prompts from scratch.

The goal isn't to keep your agents on a short leash forever. It's to build toward real autonomy through a process you can actually feel good about.

Before you go, I want to zoom out for a second.

What used to happen in a quarter is now happening in a week. Three flagship model releases, a $13 billion infrastructure deal, all in seven days. The pace is not slowing down.

Google Cloud Next is this week and Meta's LlamaCon is right around the corner on April 29. I'll be watching both and will have the key takeaways in the next edition.

One question to leave you with: of everything in this edition, what's the one thing you're actually going to try this week?

Cheers,

Julien

PS: If you know someone who would value the information I share each week please forward this to them, or send them a link to subscribe at www.ampra.ai/join-our-newsletter 

Keep Reading