The recent launch of Claude Cowork has generated real excitement in the market. It signals how quickly AI is moving from chat-based assistance into hands-on business tasks.
That excitement is warranted. It is also exactly why we decided to test Claude Cowork against real-life enterprise workflows.
We were less interested in what AI can do in a controlled demo and more interested in what happens when it is applied to the messy, high-stakes work of the enterprise: large spreadsheets, long PDFs, repeatable data preparation, multi-step processes, integration across legacy platforms, and outputs that need to stand up to review, maintenance, and governance.
That includes the kinds of workflows enterprises deal with every day — reconciliations, consolidations, accruals, sales tax, transfer pricing, and broader month-end close and tax provisioning workflows.
We found that, in many cases, AI is remarkably capable. But capability alone is not the same as operationalization. The challenge is not just getting AI to produce an answer, but turning that intelligence into a workflow that an enterprise can actually run, trust, and govern.
That is the gap we set out to examine.
Enterprise AI Evaluation Framework: Thinking, Planning, Execution

In our view, enterprise AI should be evaluated across three distinct layers: thinking, planning, and execution.
Thinking is the model’s ability to understand intent, reason through ambiguity, generate logic, and propose an answer. This is where modern AI already shines. It can interpret natural language, make sense of messy inputs, and produce surprisingly strong first-pass outputs.
Planning is the step between intelligence and action. It is the ability to translate that thinking into a repeatable, human-readable process that a business can understand and trust. A real plan should define the workflow steps, account for validation and exceptions, incorporate human review where needed, and be maintainable over time without requiring a complete rewrite.
Execution is the ability to run that plan reliably in the real world: across large files, messy documents, business systems, and recurring workflows, with the controls enterprises actually require. That includes scale, auditability, versioning, approvals, security, and governance.
The distinction between these layers matters because enterprises do not create value solely from AI thinking. They create value when that thinking can be operationalized into repeatable, reviewable, and governable workflows.
Our research suggests that AI is already strong at thinking, less reliable at planning, and still limited in terms of execution for enterprise-grade workflows.
Thinking: The Reasoning Layer
We first looked at the core reasoning layer: can the model understand business intent, interpret messy inputs, and generate a plausible approach to the task? In that dimension, Claude Cowork is genuinely impressive.
Tests

In data-heavy workflows, we tested whether the model could take natural-language instructions and translate them into meaningful business logic across spreadsheet- and analytics-oriented tasks. What we found was encouraging: Claude was typically able to handle structuring, filtering, joining, reshaping, and other common transformations from plain-English instructions. Testers repeatedly described the experience as intuitive, easy to learn, and surprisingly accessible.
In document-heavy workflows, we tested whether the model could reason through messy, unstructured inputs such as long PDFs and invoice-style documents. Here, too, the results were directionally strong. Claude could frequently identify structure, separate distinct records, and produce a reasonable first-pass extraction from messy files. In some cases, it correctly inferred that a long document contained multiple records and separated them accordingly. In others, it was able to work across very large documents and consolidate outputs into a single result.
We also tested whether the model could improve its reasoning when given feedback, answer keys, or additional instructions. That turned out to be another area of strength. In most cases, Claude could revise its output, explain what it got wrong, and improve on a second pass, making it useful not just for first-pass output, but for iterative exploration and refinement.
Observations
The common thread across these tests was that the model understood the assignment. It could interpret intent, reason through ambiguity, and get to a plausible answer quickly.
That matters more than it may seem. In many enterprise workflows, the bottleneck is rarely not knowing what needs to happen. Most often, it’s knowing how to express that logic in the syntax of legacy tools the business uses. AI meaningfully lowers that barrier and compresses the distance between business intent and technical action.
We also observed that the model was usually directionally right even when the output was incomplete. That makes it useful as a first-pass reasoning system, especially in workflows involving messy data or unstructured documents.
What We Learned
- AI is already strong at reasoning through business tasks.
Across both data-heavy and document-heavy workflows, the model showed that it can understand intent and generate plausible approaches with relatively little instruction.
- AI reduces the translation burden for business users.
Users do not need to define every step up front with complex formulas or rigid logic just to get started.
- First-pass intelligence is real value.
Even when outputs were incomplete, the ability to produce a directionally strong first pass made AI useful for exploration, prototyping, and early-stage analysis.
Planning: From Probabilistic Reasoning to Deterministic Workflows
We next looked at the planning layer: can the model turn a plausible answer into a repeatable, human-readable workflow that a business can understand, validate, and maintain over time?
Tests

We tested whether the model would produce a business-readable workflow definition, rather than just an answer or implementation artifact. What we found was that the “plan” was often not a human-readable process at all. It was code, fragments of logic, or a mix of narrative and generated scripts. In some cases, Claude created step-by-step documents or custom skills, but these did not provide a workflow definition that a team could inspect and manage.
We also tested whether the logic produced by the model could be understood and maintained over time by a business team. In spreadsheet-oriented workflows, Claude frequently dropped into custom JavaScript, Python, bash, or XML manipulation. Even when the output looked promising, the underlying logic was difficult to follow. In some cases, the code was highly opaque or hard-coded in ways that made the result fragile and difficult to maintain.
Finally, we tested whether the workflow plan could be reused natively without reconstructing context or relying on user workarounds. What we found was that preserving the process often depended on asking Claude to document what it had done, then reusing that write-up later. That may be useful, but it is not the same as having a native, repeatable workflow plan.
Observations
The core issue was not that AI had no plan. It was that the plan often lived inside the model, or inside generated code, rather than as a human-readable process the business could own.
Another way to describe the gap is this: enterprise workflows require turning probabilistic AI reasoning into deterministic business processes. A model may generate a plausible answer or path forward, but a business cannot run critical workflows on plausibility alone. The logic has to be made explicit, assumptions have to be surfaced, validation has to be built in, and the process has to become repeatable enough to produce dependable outcomes over time.
We also observed that validation, exception handling, human review points, and maintainability were not naturally part of the process unless the user explicitly forced them in. Even then, those elements were often layered on after the fact rather than designed into the workflow from the start.
What We Learned
- AI can produce a path without producing an enterprise-ready plan.
Generating steps is not the same as defining a process a business can inspect, trust, and maintain.
- AI struggles to turn probabilistic reasoning into deterministic workflow logic.
That translation is still largely left to the user.
- Human readability is essential to operationalization.
If workflow logic is embedded in opaque code or model behavior, it becomes difficult for teams to validate, modify, and reuse over time.
Execution: Trusted, Governed Outputs at Scale
We then looked at the execution layer: can the model reliably run the workflow against real-world files, systems, and business constraints, while producing outputs that can be trusted and governed at scale?
Tests

In one test, we evaluated whether the system could handle a large, business-critical spreadsheet and perform a relatively simple update without damaging the file. What we found was a meaningful breakdown in execution reliability. The output file was corrupted and could not be reliably opened. The process took far longer than the manual alternative, and substantial human effort was required to check the results and recover from the failure. In one run, the workflow became resource-intensive enough to destabilize the machine.
In another test, we looked at whether the system could extract useful information from long, messy PDFs with enough accuracy to reduce review effort. What we found was more mixed. Claude could produce a reasonable first pass and return directionally useful results. But accuracy became uneven when tables, line items, and visually structured content mattered. In some cases, answer keys, rework, and additional checking were still required before the output could be trusted.
We also tested whether the system could support repeatable execution of workflows over time. Claude could generate shortcuts, document its own steps, or attempt to create reusable skills. But these were not equivalent to a durable workflow execution layer. Re-running the process frequently required the user to return to the desktop app, reconstruct context, or use documentation as a workaround.
Finally, we tested whether the system could operate within the enterprise environment by connecting to upstream and downstream business systems. Claude was most effective when working with local files or when accessing files directly. Support for deeper connectivity to systems of record, such as ERPs, close platforms, tax engines, and downstream operational systems, was much more limited in practice. That left teams relying on manual extraction, intermediary file handling, and disconnected execution outside the core enterprise stack.
Observations
Across these tests, the pattern was consistent. The issue was not whether the AI understood the task. In many cases, it clearly did. The issue was whether it could execute reliably under real-world enterprise conditions.
We also saw that the review burden remained high. Even when the output looked plausible, users repeatedly had to spend significant time validating its correctness. That changes the nature of the value proposition: the AI may accelerate parts of the work, but it does not necessarily remove the operational burden.
Finally, many of these workflows appeared to rely on inference as the execution engine itself. That can be useful for interpretation, judgment, and extraction. But for most enterprise workflows, that is not the ideal execution model.
What We Learned
- Understanding the task is not the same as executing it reliably.
The breakdown came when workflows had to operate against large files, real business artifacts, and more demanding execution conditions.
- Inference is powerful for interpretation, but it is not ideal as the execution engine.
Most enterprise workflows would benefit from a data processing engine for scale, cost efficiency, repeatability, and deterministic output, while using AI where reasoning actually adds value.
- Enterprise execution requires an operating framework.
Task completion alone is not enough. Real execution also requires repeatability, low review burden, durable automation, system connectivity, and the surrounding framework for collaboration, control, and oversight.
Enterprise AI Workflow Scorecard
At a high level, our research points to a clear pattern: general-purpose AI is already strong at thinking, less mature at planning, and still limited at execution for enterprise workflows.
That does not diminish the value of AI. It clarifies where AI is already powerful today, and where enterprises still need an operational layer around it.

From Intelligence to Execution
General-purpose AI is already a strong reasoning layer. The enterprise challenge is turning that reasoning into a workflow that teams can run the same way every month, inspect without decoding generated code, and defend in a compliance review. Closing those gaps is what converts AI capability into the productivity gains enterprises are counting on.
Savant is built for precisely this gap. Where Claude Cowork provides strong reasoning and first-pass intelligence, Savant adds the operational layer that regulated workflows require: governed workflow definitions that teams can inspect and maintain, a data processing engine built for scale and deterministic output, and connectivity to the systems of record where enterprise finance actually lives. Together, they cover the full range from initial reasoning through trusted, repeatable execution.
See how Savant takes Claude from first-pass intelligence to governed execution across your finance organization at https://savantlabs.io/solutions/finance.