Tags
AI, Artificial Intelligence, blackboard-system, Business Rules, chatgpt, expert system, explainability, intelligent-actors, llm, MAPPER/ESD, modus-ponens, production rules, rule-set-management, rule-sets, rules, technology, Unisys
Martyn Rhisiart Jones, Sir Afilonius Rex, Lile de Alba and our agentic intelligent actor, Selina Savant.
Spain, 6th December 2025.

The Logic Layer: Why Old-School Rules Are the New Guardrails for Generative AI
A hybrid renaissance is quietly rewriting the future of trustworthy artificial intelligence. The great folks here at Good Strat humbly champion this movement. In the fevered race to build ever-larger language models, a counterintuitive truth has begun to emerge from the labs. Surprisingly, sometimes the best way to make AI smarter is to shackle it with centuries-old logic. Modus ponens is that dusty Latin phrase from Aristotelian syllogism (“if P then Q; P, therefore Q”). It is enjoying an unexpected second act. It has become one of the most promising tools for keeping generative AI honest. It helps make AI coherent and, dare one say, responsible.
Think of it as the return of symbolic AI, not to replace the neural behemoths, but to ride shotgun.
How the rule engine actually works
At its simplest, a rule-based agent is a referee sitting beside the LLM. While the language model dreams in probabilities, the logic engine speaks in certainties. Together, they create what researchers now call neuro-symbolic or hybrid architectures. These systems combine the fluid creativity of transformers with the iron rigour of formal reasoning.
The mechanics are elegant. Before an answer is served to the user, or even during generation itself, the output is run through a battery of checks:
- Consistency policing
If the model earlier stated “All men are mortal,” the rule engine throws a flag. This happens if it later implies that Socrates is immortal. Contradiction detected. - Domain-knowledge enforcement
In medicine: “Fever + neck stiffness + photophobia → meningitis must be on the differential.”
In law: “No consideration → no enforceable contract.”
Violate these encoded rules and the answer is rejected, sent back for revision, or automatically corrected. - Provenance discipline
Every claim must carry its guarantee. “Show your working,” the logic layer commands. It forces the model to trace each inference back to a premise or an evidence source. - Step-by-step compulsion
No leaping to conclusions. Each logical jump must be justified by an explicit rule. This habit dramatically cuts the rate of “reasoning hallucinations.” These are moments when an LLM confidently invents non-existent facts or invalid deductions.
The result? Outputs that are not merely fluent but verifiably coherent within the domains the rules cover.Where it shines – and where it inevitably stumbles.
These hybrid systems are already showing startling gains in tightly bounded arenas. Medical diagnostic assistants who once hallucinated rare diagnoses of zebra now stick closer to evidence-based guidelines. Legal research tools produce fewer fictional case citations. Mathematical reasoning benchmarks – once the graveyard of pure LLMs – have seen leaps ahead when chained to symbolic solvers.
Yet no one in the field pretends this is a universal solvent. Human knowledge is messy, ambiguous, and constantly evolving. You can encode the rules of chess. Similarly, you can encode the differential diagnosis of meningitis. Still, writing an exhaustive rule set for geopolitical strategy in 2025 is much more challenging. The same goes for the nuances of literary interpretation. Over-zealous rule systems can also sterilise creativity. Ask a heavily constrained model to write poetry or brainstorm moon-shot business ideas, and you’ll get safe, predictable pap. The sweet spot, researchers agree, is selective application. Unleash the full generative fire for open-ended tasks. Then bring in the logic hammer only when correctness genuinely matters.
The new architectural playbook
Three patterns dominate the cutting edge:
- Post-hoc verification
LLM generates freely → logic engine audits → approved or corrected output. - Constrained decoding
The rule engine imposes guardrails in real time, pruning improbable token paths that would violate known truths. - Multi-agent debate
One agent proposes. A second agent challenges on logical grounds. A third fact-checks the information. This is a digital version of the scientific peer-review process.
You can already glimpse these ideas in the wild:
OpenAI’s o1 series comes with hidden “test-time compute” reasoning chains. Anthropic’s Constitutional AI includes classifiers. Google’s AlphaProof system blends LLMs with formal theorem provers. There is a flourishing academic ecosystem of “self-consistency,” “tool-former,” and neuro-symbolic papers.
The bottom linePure scale is no longer enough.
As generative AI edges closer to high-stakes deployment, the pressure for verifiable correctness becomes existential. These high-stakes areas include clinical decision support, legal advice, and scientific discovery. Rule-based logic, far from being a relic of 1970s expert systems, is re-emerging as an essential seatbelt.
The future almost certainly belongs to hybrids: language models for intuition and fluency, symbolic engines for discipline and truth. Probabilistic dreaming creates an elegant tension with logical rigour. In this balance lies the next leap toward artificial intelligence we can actually trust.
And that, perhaps, is the most rational conclusion of all.
Many thanks for reading, and have a great rest of the weekend.
Tomorrow we will have a panel of guests interpret this article for your education, information and entertainment.