AI
AI cannot fix bad strategy
AI can optimise, generate, and accelerate, but it cannot correct a product direction that was flawed from the start. It amplifies the strategy it is given.
Why AI only creates value when it is anchored to a clear direction, and why using it on top of weak strategy often makes the underlying problem harder to see.
In short
Why AI only creates value when it is anchored to a clear direction, and why using it on top of weak strategy often makes the underlying problem harder to see.
Why AI feels like progress even when direction is unclear
Rewrite the content. Optimise the glossaryPain PointA specific problem or frustration users experience when trying to complete a task.Open glossary term. Personalise the experience. Let the glossaryModelA model is a system or representation used to process data and generate outputs, often trained to perform specific tasks.Open glossary term figure it out.
On the surface, that feels like progress.
But it only works if the direction is right to begin with.
Because AI does not correct glossaryStrategyStrategy is a high-level plan that defines long-term goals and the approach to achieving them.Open glossary term.
It amplifies it.
AI does not rescue unclear thinking. It simply scales whatever direction the product is already moving in.
Why most product problems start before execution
In my experience, most of the problems that show up in digital products are not caused by poor execution. They sit much earlier. The wrong problem has been prioritised. The glossaryPain PointA specific problem or frustration users experience when trying to complete a task.Open glossary term does not reflect how users actually behave. The structure does not support the task it is meant to enable. The product is solving something, just not the thing that matters most.
When that happens, no amount of glossaryOptimisationOptimisation is the process of improving a product or journey to increase performance, usability, or conversion.Open glossary term will fix it.
AI just makes it happen faster.
Key takeaway
If the problem has been framed badly, AI can improve the execution while leaving the real issue completely untouched.
Why optimisation is not the same as effectiveness
Content can be improved, but if it is answering the wrong questions, it will still miss the mark. glossaryPain PointA specific problem or frustration users experience when trying to complete a task.Open glossary term can be streamlined, but if they are built on incorrect assumptions, they will still create glossaryFrictionFriction refers to anything that slows users down or makes it harder for them to complete a task. It can be caused by poor design, unnecessary steps, unclear messaging, or technical issues.Open glossary term. glossaryInterfaceAn interface is the point of interaction between a user and a system, where inputs are made and outputs are received. It can be visual, physical, or conversational.Open glossary term can be refined, but if the underlying flow is flawed, the experience will still break down.
Everything becomes more efficient.
But not more effective.
Why AI can create a false sense of progress
This is where the risk sits.
Because AI is very good at making things look better.
It can produce cleaner copy, more consistent structures, and more polished outputs. It can remove obvious glossaryFrictionFriction refers to anything that slows users down or makes it harder for them to complete a task. It can be caused by poor design, unnecessary steps, unclear messaging, or technical issues.Open glossary term, smooth out rough edges, and create something that feels more complete. From the outside, it can look like the product has improved significantly.
But underneath that, the same issues often remain.
The glossaryStrategyStrategy is a high-level plan that defines long-term goals and the approach to achieving them.Open glossary term has not changed.
What I have seen is that this can create a false sense of progress. Metrics may shift slightly, glossaryEngagementEngagement refers to how users interact with a product, content, or experience, including actions like clicks, time spent, and interactions.Open glossary term may improve in places, and there is a sense that things are moving in the right direction. But the core problem is still there, just hidden behind better execution.
That only holds for so long.
Why users still respond to relevance, not polish
Because users do not respond to polish.
They respond to glossaryFaceted NavigationFaceted navigation is a system that allows users to filter content across multiple attributes simultaneously, combining different criteria to refine results.Open glossary term.
If the product does not align with what they need, how they think, and what they are trying to achieve, no amount of glossaryRefinementRefinement is the process of preparing and clarifying backlog items before development.Open glossary term will compensate for that. They may spend a little more time, click a little further, or engage slightly more, but eventually the gap becomes clear.
And they glossaryDrop-offDrop-off refers to users leaving a journey before completing a desired action or reaching the next step.Open glossary term.
Why strategy has to come before AI
This is where glossaryStrategyStrategy is a high-level plan that defines long-term goals and the approach to achieving them.Open glossary term needs to come first.
Before AI is introduced, before glossaryOptimisationOptimisation is the process of improving a product or journey to increase performance, usability, or conversion.Open glossary term begins, before anything is generated, the direction needs to be clear. What problem is being solved? Why does it matter? How does the product support that in a way that makes sense for the user and the business?
Without that glossaryClarityClarity is how easily users can understand what is happening and what they need to do.Open glossary term, AI has nothing solid to work with.
It is just filling in gaps.
What AI is genuinely powerful at once the strategy is sound
Used properly, AI can be incredibly effective once the glossaryStrategyStrategy is a high-level plan that defines long-term goals and the approach to achieving them.Open glossary term is right. It can scale content, accelerate glossaryIterationIteration is the process of repeatedly improving a product through cycles of testing, feedback, and refinement.Open glossary term, and support glossaryRefinementRefinement is the process of preparing and clarifying backlog items before development.Open glossary term in a way that would otherwise take significantly more time. It can help explore different approaches and surface opportunities that might not have been obvious.
But it needs something to anchor to.
That anchor is glossaryStrategyStrategy is a high-level plan that defines long-term goals and the approach to achieving them.Open glossary term.
When the direction is clear, AI becomes a multiplier. It strengthens what is already working, helps move faster, and supports better outcomes. When the direction is wrong, it simply reinforces the problem, making it harder to recognise and more expensive to fix later.
Why AI is not a safety net for poor thinking
This is the distinction that matters.
AI is not a safety net.
It is not there to correct poor thinking or compensate for unclear direction. It will not step in and identify that the wrong problem is being solved. It will not realign a product with user needs if that glossaryAlignmentAlignment is the shared understanding and agreement between teams, stakeholders, and objectives.Open glossary term was never there in the first place.
That still requires human judgement.
The temptation is to use AI to improve what exists.
The reality is that sometimes what exists needs to be challenged first.
Until that happens, no tool will fix it.
AI cannot rescue a bad glossaryStrategyStrategy is a high-level plan that defines long-term goals and the approach to achieving them.Open glossary term.
It can only execute it more efficiently.