top of page

The New Claude 4 Can Code, But Leaders Should Still Sign Off

Claude 4 is a leap forward, but it's also a governance wake-up call

By Tommy Cooke, powered by caffeine and curiousity

May 30, 2025

Key Points:


  1. Delegating a technical task doesn't guarantee it's done right—oversight matters, even when the system looks competent


  2. Claude 4’s ability to work autonomously highlights the growing need for clear accountability and human verification


  3. As AI systems become more capable, leaders must stay close to the outcomes—even when they don’t touch the inputs


In my formative years as a young adult, I was an active musician in a rock band. My band and I performed regularly throughout my undergraduate years. As much fun as gigging is, live shows are a scramble. Hauling gear, setting up, sound-checking, and hoping nothing went wrong.


At one show, I was behind the ball a bit. Two strings snapped on my main guitar. So, I asked the venue’s sound tech to wire up my pedalboard while I handled other setup tasks. When we hit soundcheck, my sound was a mess.


One of the pedals had been placed in the wrong order. It was a simple mistake, but it could have derailed the entire show. I’ve never forgotten the lesson that delegating a technical task doesn’t mean it’s done right. You still need to check the signal before the lights go up.


This is a moment I reflected on when I read that Anthropic released Claude 4. The reflection was triggered by the fact that most headlines focused on one detail: the model can autonomously generate software for hours at a time. For developers, this is surely a turning point. An AI system that not only writes code but improves it quietly, efficiently, and without supervision.


But that’s not the full story. If you are a Pro, Max, Team, or Enterprise Claude plan user, this matters to you because you will have access to Claude 4. This means you and potentially your organization may now have access to a brand new, advanced AI that can carry out complex work without human input. It means that leadership must ask: what’s the governance plan? Who verifies the output? Who signs off?


Much like the way in which I learned from asking someone who doesn’t know my system to essentially set it up for me on a rock stage, there’s something we as business leaders can do to ensure that AI innovations still act effectively on our behalf.


Claude 4 and the Shift to Autonomous Execution


Until recently, generative AI required heavy user input. The human wrote the prompt, and the system responded. That dynamic made it easy to keep the human in the loop to control the task, validate the output, and decide what comes next.


Claude 4 changes these terms. It introduces what many call agentic AI: models capable of reasoning through tasks, planning multi-step actions, and executing work without continual prompting. Claude 4 is demonstrating that it can work independently for hours, reconfigure code, and make judgment calls along the way. So, it’s not just writing the code—it is actually finishing the job as well.


This is a major development. But with this innovation in AI autonomy comes a truth: the more work that AI performs alone, the less visibility organizations have into how it gets done.


The AI Governance Gap Is Growing


The risk isn't that AI will make obvious errors. It’s that it will produce plausible work that quietly deviates from your standards, assumptions, or intentions. Do you have someone in place that can notice these changes before it’s too late? That’s the real governance gap. It’s not about control over prompts, but rather prioritizing oversight over outcomes.


This means organizations need to reconsider how they monitor AI-driven work. That doesn’t mean leaders need to personally review every AI-generated output. But it does mean they need to put in place clear lines of accountability, regular review processes, and internal checks to ensure AI isn’t working in a vacuum.

This also means that oversight can no longer be reactive—it needs to be built in from the beginning.


What Does Accountable AI Adoption Look Like?


Organizations don’t need to halt progress to manage these risks, but they do need to move forward with clarity:


  1. One of the most effective ways to begin is by documenting how AI is being used across the business. This doesn’t need to be a heavy-handed process. Even a lightweight registry of AI use cases can help identify where autonomy is increasing and where review protocols might be missing


  2. Leaders should also establish guidelines for when human oversight is required. Not every AI-generated output requires manual review, but some certainly do. Defining these boundaries in advance protects against over-reliance on unchecked systems


  3. Lastly, every autonomous system should have a clearly named owner. Someone in the organization needs to be responsible for verifying that the AI’s work aligns with business objectives, ethical expectations, and legal obligations. The idea isn’t to create bottlenecks—it’s to make sure someone is watching.


Signing Off Is Still a Human Task


Claude 4 marks real progress. It moves us closer to a world where AI can take on meaningful work, save time, and support innovation. But that progress also demands more from leadership.


Delegating work to machines doesn’t absolve humans of responsibility. If anything, it raises the bar because the more invisible the work becomes, the more deliberate our oversight must be. Leaders don’t need to fear these systems. But they do need to govern them. They need to understand where AI is being used, what it’s allowed to do, and who remains accountable when things go wrong.


This type of oversight can help organizations explain how their AI systems generate outputs.

bottom of page