Exploring How Deterministic Systems and LLMs Should Actually Work Together.
I’ve spent 15+ years building the data pipelines that run businesses — SQL, VBA, Python, SSIS, Power BI. Now I’m working through how LLMs fit into that same continuum: where they belong, where they don’t, and what it costs. This site is where I share that thinking as it evolves.
Where Things Actually Stand
There’s a significant gap between what organizations expect from AI and what they’re actually prepared to implement. The data is striking — and it consistently points to the same root causes.
of enterprise AI projects never reach production. Harvard Business Review reports this is nearly double the failure rate of traditional IT projects — and the trend is accelerating, not improving.
of the obstacles come from people and processes, not the technology itself. BCG found that only 10% of AI failures are actually about the algorithm. The rest is organizational: governance, data readiness, architectural clarity.
of enterprises say their data isn’t clean or reliable enough for AI. Meanwhile, 84% of practitioners encounter conflicting versions of the same metric. AI doesn’t solve that gap — it exposes it.
of organizations lack a clear measure of AI value. Half of CEOs believe their job depends on getting AI right — but without cost visibility at the token level, there’s no way to know if it’s working.
Sources: Harvard Business Review (2023), BCG “From Potential to Profit” (2025), Modern Data Report 2026, S&P Global Enterprise AI Survey (2025), BCG AI Radar via WEF (2026), MIT Project NANDA (2025), Deloitte AI Token Economics (2026).
The Numerate Semantic Control Plane
I haven’t seen a great implementation framework for integrating LLMs into enterprise data workflows yet — one that treats them as what they are (word predictors) and keeps everything else deterministic. So I’m building one. The core idea: a semantic layer owns all business definitions and calculation rules, deterministic engines handle computation, and LLM nodes handle only what code cannot — intent classification, exception routing, narrative generation. Every output traces to source data, every action passes through compliance controls, and every inference call is tracked by cost. This is a work in progress. The scenarios below use finance as the example domain, but the five-layer pattern is designed to adapt to any regulated environment.
| Metric ID | Name | Category | Formula / SQL | Data Products | Target |
|---|
| ID | Control Name | Type | Frequency | Owner | Data Products | Status |
|---|
Token Economics
If AI costs are consumption-based and unpredictable, organizations need new tools to think through them. This is an evolving toolkit for exactly that: model pricing comparisons, inference gateway architecture, ROI modeling, and optimization strategies — grounded in current provider rates and real cost structures. Deloitte calls tokens “the new currency of AI.” This section explores what that means in practice.
| Model | Provider | Input $/1M | Output $/1M | Context | Tier | Best For |
|---|---|---|---|---|---|---|
| Claude Sonnet 4.6 | Anthropic | $3.00 | $15.00 | 200K | 1 | Financial analysis, compliance docs |
| GPT-5.2 | OpenAI | $1.75 | $14.00 | 200K | 1 | Complex reasoning, exception triage |
| GPT-5 Mini | OpenAI | $0.25 | $2.00 | 200K | 2 | Forecasting, high-volume tasks |
| Claude Haiku 4.5 | Anthropic | $1.00 | $5.00 | 200K | 2 | Budget checks, routing |
| Gemini 2.5 Flash | $0.30 | $2.50 | 1M | 2 | High-volume ad-hoc queries | |
| DeepSeek V3.2 | DeepSeek | $0.28 | $0.42 | 128K | 2 | Classification, extraction |
| GPT-5 Nano | OpenAI | $0.05 | $0.40 | 400K | 2 | High-volume simple tasks |
Initiates call
Gateway
Auth + Route
Budget + Token Est.
Tier enforcement
Provider
Receives result
Strip metadata
Actuals + Budget
+ Usage headers
Availability: 99.95%
Peak RPS: 247
Cache hit: 44.2%
Payback assumes estimated platform setup cost ($5K–$60K based on volume)
| Optimization | Applicable Calls | Reduction | Monthly Savings | Effort |
|---|---|---|---|---|
| Model Routing | 70% | 65% | $812 | Low |
| Prompt Caching | 44% hit | 70% | $218 | Med |
| Prompt Engineering | 100% | 20% | $110 | Med |
| Batch Processing | 35% | 25% | $64 | High |
| Semantic Caching | 15% | 100% | $47 | High |
| Total (combined) | ~80% | $1,251/mo | ROI: 12x |
The Path Here
From early web development through financial domain depth to enterprise systems leadership — I’ve built my career at the convergence of finance, data, and technology.
This background means I speak finance and engineering fluently — the rare overlap that lets me diagnose in days what keeps departments stuck for quarters.
What I Work On
Finance Data Architecture
Governed, auditable data architectures — from ERP integration through warehouse design to compliance-ready pipelines.
- Enterprise GL, AP, AR, and reconciliation data platforms
- ERP integration, ETL pipelines, data warehouse design
- Data governance, quality frameworks, SOX-compliant architectures
BI & Analytics Engineering
Dashboards, reporting, and analytics platforms that turn raw data into decisions — built by someone who understands the underlying business logic.
- Executive dashboards, self-service analytics, operational reporting
- Power BI, Tableau, SSRS development and optimization
- KPI frameworks, metric catalogs, semantic layers
AI-Ready Architecture & Governance
Forward-looking architecture thinking for how LLMs integrate into enterprise workflows — with the governance frameworks and cost visibility that production deployment will require.
- Agentic workflow design — LLMs as semantic routers, not calculators
- Token economics and inference cost tracking — the metrics executives will need before approving any AI workflow
- The Numerate Semantic Control Plane — a reference architecture demonstrated below
- Governance-first approach: audit trails, human approvals, regulatory-compatible controls
Let’s Talk
I’m sharing these frameworks and tools because I think the conversation around enterprise AI needs more architectural rigor and less hype. If any of this resonates — or if you see gaps I should address — I’d welcome the conversation.