Categories
Blog

AI-Assisted Decision-Making in Regulated, High-Stakes Environments

In this piece
  • why the main constraint in serious AI systems is usually the operating environment, not the model
  • where AI initiatives fail in regulated and high-stakes settings
  • what leaders should ask instead if they want usable, governable systems

The value of AI in regulated and high-stakes environments is not primarily in generation. It is in improving judgment.

That is the core point.

In serious operating environments, AI becomes useful when it helps people find the right information, frame the right decision, and move with more clarity under real constraints. It becomes dangerous when it is treated as a shortcut around operating discipline.

That is why the usual public conversation about AI still misses the point. It is too focused on the model and not focused enough on the surrounding system.

The Real Constraint Is Not the Model

Most AI commentary assumes the working environment is clean, reversible, and low-stakes.

It assumes the documents are usable, the metadata is coherent, the workflow is simple, and the consequences of error are limited. In that frame, the main question becomes whether the model is capable enough.

That is not how serious environments work.

Once the stakes rise, the limiting factor is rarely the raw model. The limiting factor is the system around it:

  • the quality of the underlying documents
  • the reliability of metadata and retrieval
  • the fit with real operational workflows
  • the level of governance and traceability required
  • the ability to defend decisions under scrutiny

That is where many AI initiatives fail. The demo works. The operating system does not.

What AI Is Actually Good For

In regulated and high-accountability settings, AI is most useful when it strengthens five things.

1. Faster access to relevant information

In document-heavy environments, the first problem is often not analysis. It is access. Teams need to find the right material quickly, even when the information estate is large, fragmented, and inconsistently structured.

2. Better framing of the problem

AI can help surface patterns, summarize complexity, and expose gaps. That matters because many bad decisions start with a badly framed problem.

3. Stronger decision support

Used properly, AI can help compare options, clarify constraints, and support faster judgment without pretending to replace judgment.

4. Operational leverage

Well-designed AI systems can reduce friction in repetitive analysis, information triage, and early-stage interpretation. That is less glamorous than the market narrative, but much closer to where the practical value sits.

5. More usable systems under pressure

The real test is whether the system still works when the environment is messy, time is limited, and the decision matters.

Where Things Break

Most failures are not model failures. They are operating failures.

Documents are often low quality. Some are structured and usable. Others are scanned, inconsistent, or broken by poor formatting. Tables do not parse properly. Context gets split across appendices, attachments, and disconnected files.

Metadata is often weaker than teams believe. Many organizations think they have a searchable knowledge base when they actually have a large archive with unreliable naming, weak tagging, and inconsistent structure.

Workflow fit is another common failure point. A tool may perform well in isolation but fail once it has to sit inside real review, governance, or decision-making processes. If it cannot survive handoff, scrutiny, and audit pressure, it is not operationally useful.

Then there is governance. In regulated environments, traceability is not optional. Evidence matters. Decision logic matters. The ability to explain how a conclusion was reached matters. If AI increases speed but reduces defensibility, it weakens the system rather than improving it.

Why Regulated Environments Matter

Regulated environments reveal the truth about AI faster than most other settings.

That is one reason I find them so useful as a lens.

In MedTech, diagnostics, laboratory services, and other high-accountability contexts, the cost of sloppy thinking is higher. Decisions affect compliance, commercial risk, operational credibility, and sometimes patient pathways. Weak systems get exposed quickly.

That pressure forces better questions:

  • Can the information be trusted?
  • Can the process be repeated?
  • Can the output be governed?
  • Can the workflow survive real-world use, not just a pilot?
  • Can the organization explain what it is doing if challenged?

These are not anti-AI questions. They are the questions that make AI useful.

The Operating System Matters More Than the Demo

The real opportunity is not in producing more impressive demos. It is in building systems where AI sits inside a disciplined operating structure.

That means:

  • better document handling
  • stronger metadata design
  • clearer decision pathways
  • tighter workflow integration
  • appropriate governance and traceability

The model matters. But in serious environments it is rarely the decisive factor. The decisive factor is whether the surrounding operating system makes that intelligence usable, governable, and reliable.

What Leaders Should Ask Instead

If you are evaluating AI for a regulated or high-stakes environment, the most useful questions are not the usual ones.

Not:

  • which model is newest
  • which demo looks most impressive
  • which vendor promises the most automation

Instead:

  • what decision-quality problem are we trying to improve?
  • how good is the underlying information base?
  • where do document quality or metadata failures break the workflow?
  • what level of traceability do we need?
  • what would make this system trustworthy in day-to-day use?
  • where does human judgment remain essential?

Those questions are less marketable than the standard AI pitch. They are also much closer to the truth.

Closing

AI becomes more valuable as the consequences become clearer.

That is where weak systems fail and where well-designed systems create real leverage. In regulated and high-stakes environments, the future of AI will be decided less by novelty and more by whether the surrounding system is strong enough to make that intelligence useful under real pressure.

If you are working through that kind of challenge, explore the related essays or get in touch.

References and Frameworks

  • Barbara Minto’s Pyramid Principle as a useful reference for top-down communication
  • the practical distinction between model capability and operating-system strength
  • regulated-environment design principles: traceability, defensibility, and workflow fit
Categories
Blog

From LLMs to Agentic Workflows: How Domain Intelligence Matures

In this piece
  • why fluent LLM output is not enough in regulated or expert domains
  • how AI systems mature from basic prompting to governed agentic workflows
  • when risk, traceability, and scrutiny make more structured systems necessary

Large language models changed how people interact with information. They can summarise, explain, draft, and reason at a level that would have seemed implausible only a few years ago.

But once AI moves into domain-specific, high-stakes contexts such as regulation, law, finance, healthcare, or strategy, fluency stops being enough. The relevant question becomes simpler and more demanding: when is an LLM enough, and when do you need something more governed?

This piece explains the progression from basic LLM usage to agentic workflows, and why that progression matters when the answer has to stand up to scrutiny.

The Core Problem: Fluency Is Not Reliability

LLMs are excellent at producing answers that sound correct. That strength is also their weakness.

In general domains, that trade-off is often acceptable. In regulated or expert domains, it is not.

Domain work usually requires four things at once:

  • correct interpretation of formal rules
  • clear jurisdictional boundaries
  • traceability back to authoritative sources
  • defensible reasoning under scrutiny

LLMs alone do not guarantee those properties.

A Maturity Continuum, Not a Binary Choice

Applied AI in serious domains usually evolves along a continuum rather than a clean split between “chatbot” and “agent”.

1. Basic LLM

General reasoning and language generation. Fast, flexible, and useful, but ungrounded.

2. LLM + Prompt Discipline

More consistency through structured prompts, but still heavily reliant on model recall and inference.

3. LLM + Retrieval-Augmented Generation

Answers are grounded in documents, policies, or knowledge bases instead of model memory alone.

4. Semi-Agentic Systems

Tools, checks, and limited validation steps are introduced to improve control.

5. Full Agentic Workflows

Explicit rules, verification steps, jurisdiction control, and failure handling become part of the operating system.

Each move to the right adds control, reliability, and auditability. Agentic workflows are not an alternative to LLMs. They are the governed, operational form of using them when the stakes are real.

When Agentic Becomes Necessary

The deciding factor is not technical sophistication. It is risk.

Two questions usually determine the appropriate architecture:

  • what is the cost of being wrong?
  • do I need to explain or defend the answer to someone else?

If both are low, a simple LLM may be sufficient. If either is high, relying on a single model becomes dangerous.

That is why agentic workflows appear first in regulatory analysis, legal reasoning, financial decision support, and safety-critical or compliance-driven domains. In those contexts, confidence without justification becomes a liability.

What Makes an Agentic Workflow Different

An agentic workflow introduces elements that LLMs do not provide on their own:

  • explicit rules encoded outside the model
  • source authority with clear prioritisation of documents, clauses, or standards
  • verification steps before answers are finalised
  • failure states so the system can stop, flag uncertainty, or request clarification
  • traceability from conclusion back to inputs and rules

This is the difference between a conversational assistant and a domain system.

Why the Extra Complexity Can Be Worth It

Agentic systems are more complex to build. That complexity only creates value when it buys certainty.

In low-risk scenarios, complexity is wasteful. In high-risk scenarios, simplicity can be irresponsible. The mistake many organisations make is treating all AI use cases as equal. They are not.

A Practical Rule of Thumb

  • if an answer only needs to help you think, use an LLM
  • if an answer must stand up to scrutiny, use an agentic workflow

Closing

The future of applied AI is not about choosing between LLMs and agents. It is about placing LLMs inside systems that understand rules, risk, and responsibility.

That is how domain intelligence matures.

Infographic

Categories
Blog

Making AI Work for You, Not Against You

In this piece
  • why deterministic workflows usually create more value than vague prompting
  • how context, format, and scope control improve the quality of AI output
  • why verification matters more than benchmark scores in real-world use

According to ChatGPT, I was in the top 1% of its worldwide users in 2025. Its own description of my profile was blunt: this was power-user, professional-grade usage built around thinking, drafting, and systems work rather than entertainment.

That is less interesting as a statistic than as a prompt for reflection. After a year of heavy use, some patterns became clear. These are the lessons that mattered most.

Lessons From a Year of Serious AI Use

1. Deterministic beats vague

The closer you are to working in a deterministic rather than probabilistic way, the better results you get. Set the rules up front so the model knows how to behave. Personalise it where memory is available. Consistency compounds when you want to work at speed.

2. Start with the end goal

Explain the context and the outcome you want. Is the model helping you think, review, generate possibilities, summarise, or explain? The better the goal is defined, the more useful the output becomes.

3. Match the form to the task

Think about the form of the answer before you ask for it. Some problems are better handled as a table, some as a short framework, some as bullets, and some as prose. The wrong output shape often creates avoidable confusion.

4. Tables often improve thinking

A lot of output becomes easier to understand in table form. Tables help compare dimensions, narrow scope, and reduce the chance of drifting into loosely connected ideas.

5. Scope creep is real

AI will often try to guess beyond what you asked. That can feel helpful, but it can also introduce embellishment before the primary question is answered. Keep the task bounded.

6. Benchmarks do not equal trust

Most AI benchmarks do not reflect real-world use. A model can score well and still fail at something basic such as getting a reference right. For important work, verification matters more than the benchmark headline.

7. Long conversations drift

Long-form conversations become less accurate over time. Alongside scope creep, there is coherence creep. The longer the thread, the greater the risk of losing the original frame or corrupting the memory trail.

8. Start broad, then go narrow

When complexity is high, begin with architecture and model first, then move into instructions and detailed tasks. It helps to agree the high-level map before pushing into execution.

9. Ask for feedback on your own use

After enough usage, your working style becomes visible to the model. That makes it possible to ask for feedback on how to improve your prompts, your framing, and your questions.

Infographic