Over the past two years since Copilot’s rollout, I have seen a growing pattern in how teams talk about AI in Excel.
The question is almost always framed along these lines:
- “How can we use Copilot for this?”
- “Can AI replace this workflow?”
- “Should we be using AI here?”
What is striking is that these questions usually come after a tool has already been chosen, not before. AI shows up as a solution looking for a problem. And that is where teams start getting into trouble.
The real issue is not whether AI is powerful. It clearly is. The issue is whether AI is appropriate for a given task. Excel already has a mature ecosystem of deterministic tools like Power Query, formulas, PivotTables, Power Pivot and even Python in Excel. Those tools exist for a reason. They are stable, repeatable, auditable, and boring in exactly the way production work should be.
AI changes the nature of work by introducing variability. Sometimes that variability is a feature. Sometimes it is a liability. The difference matters.
This post introduces a simple way to think about that difference. It starts with a diagnostic lens I call the SAFE framework, and then moves into a practical decision matrix for choosing the right tool in Excel.
AI is not a replacement for judgment
One of the biggest mistakes teams make is treating AI as a drop-in replacement for existing Excel workflows. Monthly reporting, reconciliations, KPI calculations, and regulatory metrics are often held up as candidates for automation through prompting.
That instinct is understandable, but misguided.
These workflows exist precisely because they demand consistency. The same inputs should produce the same outputs every time. Changes should be deliberate, documented, and reviewable. AI, by design, does not guarantee that. Even when outputs look stable, the underlying process is probabilistic.
Before deciding whether AI belongs in a workflow, we need a way to evaluate the task itself. That is where SAFE comes in.
The SAFE framework

SAFE is not a checklist or a process. It is a lens. It asks four questions that determine how much freedom a task can tolerate.
- Stability asks how consistent the output must be. If the result needs to be repeatable and auditable, the tolerance for AI-driven variability is low.
- Ambiguity asks how well defined the task is. Exploratory questions benefit from flexible reasoning. Well-specified procedures usually do not.
- Frequency asks whether the task is a one-off or recurring. Prompting is not automation. If something runs every week or every month, it should eventually be encoded in a deterministic workflow.
- Exposure asks who depends on the output. The more visible the result, the higher the cost of inconsistency or error.
These four dimensions do not give you a yes or no answer on their own. What they do is surface risk. They tell you whether AI usage increases or decreases confidence in the outcome.
In general, AI fits best when ambiguity is high, stability requirements are low, frequency is limited, and exposure is contained. As those conditions reverse, deterministic tools should take over.
SAFE helps you decide whether AI should even be considered. It does not yet tell you which tool to use. For that, we need one more step.
From diagnosis to decision
Once a task passes the SAFE lens, the next question becomes practical. If AI is acceptable here, where exactly does it belong in the Excel workflow?

This is where a simple two by two matrix becomes useful. The matrix is based on two dimensions.
The horizontal axis is ambiguity, ranging from well defined to exploratory. The vertical axis is stability required, ranging from low to high.
High stability, low ambiguity: Rules must hold
This is the domain of classic Excel work. Power Query, formulas, PivotTables, and data models live here for a reason. These workflows support recurring reporting, data cleanup, joins and merges, and KPI calculations. They are designed to refresh reliably. When inputs change, outputs change predictably. That is a feature, not a limitation.
Using AI here adds risk without adding much value. If a workflow refreshes, it should not require prompting. If leadership depends on the numbers, rules must hold.
High stability, high ambiguity: The judgment zone
This quadrant is subtle and often misunderstood. The outputs matter, but the rules are not fully settled.
Defining metrics, choosing KPIs, setting business rules, and interpreting edge cases all fall here. These decisions require human judgment. Excel formulas are still useful, but they serve the thinking rather than replace it.
AI can play a supporting role, but only lightly. It can help explore alternatives or sanity check reasoning. It should not be the authority. This is a human-in-the-loop zone by design.
Low stability, high ambiguity: The AI sweet spot
This is where AI genuinely works best. Exploratory analysis, early-stage investigation, draft charts, executive summaries, and sense-making all benefit from flexible reasoning. In this zone, variability is acceptable because the goal is understanding, not finality.
Copilot and Python in Excel work well here because they accelerate thinking. They help you ask better questions, see patterns faster, and communicate insights more clearly. The outputs are drafts, not deliverables.
AI belongs upstream in the analytical process. This is what that actually looks like.
Low stability, low ambiguity: The convenience zone
Not everything needs to be engineered. One-off transformations, ad hoc checks, and throwaway analysis live here. Either AI or traditional tools can work. The primary goal is speed. Overengineering is just as wasteful as overprompting.
This quadrant exists to prevent perfectionism. It gives teams permission to move on.
Why this framing matters
When AI is pushed into high-stability, high-exposure workflows, teams experience subtle failures. Results become harder to reproduce. Explanations become weaker. Trust erodes slowly.
When AI is used upstream, where questions are still forming, it has the opposite effect. It accelerates insight, improves communication, and frees analysts to focus on judgment rather than mechanics.
SAFE evaluates the task, and the matrix chooses the tool. This separation is what keeps AI useful instead of dangerous.
A final rule of thumb
If leadership depends on the output, AI stays upstream.
That does not mean AI has no role. It means its role is to support thinking, not replace production systems that already work. The teams that benefit most from AI are the ones who understand where not to use it.
If this framework resonates, the next step is seeing how it applies to your Excel workflows, reporting cycles, and decision context. I offer an initial discovery call where we walk through how my frameworks and services are typically used to help teams:
- clarify which Excel workflows should remain fully deterministic
- identify where AI can responsibly support exploration and analysis
- reduce hidden risk in reporting, metrics, and leadership-facing outputs
The goal of the conversation is to understand your context and determine whether my training, advisory, or implementation services are the right fit for your organization.
You can book a time here:
fsdfdsfasd
