I started where most operations people stop — at the edge of what the data could tell me. The tracker existed. The dashboard existed. What didn’t exist was anything that connected them to a question worth asking in real time.
So I built it. Not with a team, not with IT support, not with a budget. With a terminal, an AI collaborator, and a specific enough understanding of the operation that I could tell the model exactly where the gaps were.
The Lattice
Each node is a unit the system decided to keep — a coefficient, a conditional, a boundary that held under pressure. They accumulate across sessions, individually weightless, collectively the topology that defines what the model can reach.
Edges form where cosine similarity between embedding vectors crosses the retrieval threshold. Not declared — computed. Two facts become adjacent when their learned representations land close enough in the latent space that the graph, given a live query, pulls activation through both. That proximity is what you’re seeing rendered as light.
When a RAG query arrives mid-sort, the system traverses this graph: embedding the query, scoring similarity against every stored node, propagating relevance across the neighborhood, and surfacing the subgraph most likely to ground the answer in operational truth rather than model prior. The walk isn’t random. It converges.
What you are watching is the becoming of understanding. The moment information holds shape.
The System
How I Work
I name the exact problem. Not “the data is unreliable” — but “scanner ID substitution corrupts iGate PPH silently, and there’s no surface indicator unless you cross-reference the SOR employee summary against the sort window to see who was where and for how long.” That’s what makes solutions exact rather than approximate.
I don’t stop at solving. I build infrastructure around the solution so the next one is faster. The tracker needed a corpus. The corpus needed analysis tooling. The analysis needed a research framework. The framework needed a quality system. Each layer compounds — the system now is the product of that, not of any single build decision.
The AI collaboration is structured. Post-mortem quality scoring after every milestone. Mandatory self-backtracking before any fix. Persistent memory so decisions from six weeks ago are still in context. I didn’t build an autocomplete. I built a thinking partner with accountability, continuity, and a defined role in the research program.
What I've Built
Experience & Credentials
- Identified a critical gap in real-time operational intelligence and independently designed, built, and deployed a full analytics suite — no IT support, no administrative access, no external guidance — using AI-assisted development throughout.
- Applied generative AI as a structured development partner: multi-session workflows with post-mortem quality scoring, a custom AI memory system (SQLite + FTS5 + LangGraph) for persistent context, and directed AI to resolve every infrastructure blocker independently.
- Delivered the Hub Operations Suite end-to-end: Tracker (v1.0→v4.0, 40+ versions, 6,200+ lines), KPI Dashboard, and a multi-iframe Container application. Full lifecycle: problem identification → design → build → test → deploy → iterate.
- Built a 613-sort operational corpus (81 MB, 7 data layers per sort) and embedded 6 OR models: M/M/c queue for sort load modeling, LP-Lite staffing optimizer, κ(DOW) zone coefficients, γ weekly decay, Phase Engine, and Destination Load Intelligence.
- Engineered a multi-source data pipeline reconciling four UPS operational systems — SOR/TMS, iGate/ESR, SEAS, and CURE — resolving cross-source identity conflicts and exposing data fidelity gaps traceable to specific operational behaviors.
- Derived and validated four structural operational constants (κ, γ, ε, ρ) from the corpus; formalized results into a 97-page self-referential research monograph spanning five analytical branches — all proofs and derivations inline, no external citations for the math.
- Manages department resources for maximum output, accuracy, and efficiency; delivers orientation, skills training, and coaching; implements personal development plans and evaluations for department personnel.
- Ran neutron-star simulations on supercomputer cluster at UTEP — arxiv.org/abs/1208.4793v1
- Optimized SEM tip-building procedure at UT Austin, raising construction success rate from 78% to 94% (+20.5% relative).
How It Was Built
A technical narrative of the 0xStrategies program — from the first Excel tracker through the research program, the memory architecture, and the local intelligence layer. Written in sequence. All version numbers and dates are sourced from verified records.
It started with a practical problem. As an area supervisor for Zone 3 (PD-09 through PD-12) at the CHEMA Twilight Hub, I had already built a personal tracker in Excel — a tool for monitoring belt-level PPH, managing borrow/loan events, and keeping a running picture of my zone's performance through a four-hour sort. The tracker worked. I shared it with peers so the operation could use it collectively. But it had a ceiling: it couldn't carry history between sorts, and as the data grew, PowerQuery started hitting storage and latency walls. The tool I needed couldn't be built inside the tool I was using.
The move to HTML was pragmatic. A browser doesn't have storage ceilings. JavaScript doesn't stall on pivot recalculation. IndexedDB can hold a sort history indefinitely. What started as a platform migration quickly became something larger — because once you remove the constraints of the spreadsheet, the question of what you can build changes.
On April 18, 2026, the first HTML build began. A conversation with my operations manager about producing a post-sort report crystallized the direction: rather than a static report, build a live dashboard — something that could surface the KPIs that matter before, during, and after the sort, in one view, without opening five different exports manually. That idea and the tracker idea merged immediately. They were the same problem from two angles: one asking "what is happening right now," the other asking "how does tonight compare to every night before it."
The first deployed tool was the KPI Presort Dashboard (April 28, 2026) — eight KPI modules reconciling iGate ESR scan data, Schedule of Record headcount, CURE destination load files, and SEAS service exception reports. It didn't just display numbers. It cross-referenced them. A volume figure that didn't reconcile between SOR and SEAS was a fidelity signal, not a display artifact. That distinction — between a number and what the number means — became the architectural premise of everything that followed.
The tracker evolved from the same codebase as the dashboard, growing from a real-time belt monitor into a full analytical engine. Version 1 read iGate ESR exports and displayed belt-level PPH with conditional color formatting. By version 4.0, it had six embedded OR models, a four-phase operational model, a DOP calculator, a Timeline module, a History module with cross-sort trends, and IDB v4 persistence keeping every sort in local storage indefinitely.
Forty-plus versions across roughly three weeks. Each version was a response to something that broke in operation: a null reference that silently crashed the project renderer, a scroll-guard that let wind-down PPH contaminate the diagnostic baseline, an LP-Lite that overcapitalized available labor. The self-backtrack discipline — no fix without a named root cause first — kept each iteration honest. The architecture memo now serves as a living map of where things live and why they're there.
The dashboard and tracker were separate HTML files, opened separately, maintained separately. The Container solved the operational friction of that split by wrapping both into a single iframe architecture: one URL, both tools, no installation, no sync issues between versions. The base64 embedding approach meant the container could carry both tools as self-contained payloads — a deployment decision that required no IT ticket, no server, no dependency management. You open one file and the operation is there.
The Container reached v4.9. The rename from "capsule" to "container" (v3.2) reflected a maturation in how I was thinking about the architecture — not a single artifact, but a runtime environment with its own footer, its own versioning, and its own responsibility boundary separate from the tools it hosted.
The tracker answered questions about tonight. Master Trend Analysis was built to answer questions about structure — what patterns hold across sorts, what changes over time, what can be predicted before the sort begins. The tool runs on a 620-sort corpus and has eight tabs: prediction, zone decomposition, OR model integration, borrow/loan analytics, trend forecasting, and freight-profile analysis. It is the closest thing to a district-level analytics layer built from a single facility's data.
The corpus itself — 81 MB of unified sort history spanning December 2023 through May 2026 — is the foundation. Twenty-nine of thirty validation sorts reconcile between SOR volume and SEAS totals within rounding tolerance. Three sorts exhibit scan efficiency above 100%, which disproved an assumption I had been carrying about iGate data bounds. Correcting that assumption cascaded through every downstream fidelity calculation. That kind of correction — finding that a foundational assumption was wrong from the data itself — is what makes the corpus valuable beyond any individual query against it.
At some point the operational questions became research questions. Not because the operation stopped mattering, but because the patterns being uncovered — zone load coefficients, PPH decay constants, fidelity distortion mechanisms, the three-layer scan attribution chain — were general enough to document formally. Four research branches emerged and developed in parallel: Branch A (Hub Operations Mathematical Framework), Branch B (Hub Operations Digital Twin), Branch C (Purposeful Systems in Hub Operations), and Branch D (Unified Framework for Hub Operations Analytics), plus Sort Intelligence as a standalone track.
Each branch is internally consistent, cites verifiable operational data, and carries honest confidence intervals. The self-backtrack discipline from the tracker development carried into the research: 38 documented corrections across the program, each with a named failure mode before any revision. The most significant involved the loader scan attribution chain — a correction that changed the mechanistic claim in every branch simultaneously and required synchronized updates across all five documents in one session.
The four-branch research structure is not an organizational convenience. Mathematically, it behaves like a fixed-point iteration: each branch refines the shared model until convergence, with cross-branch corrections propagating forward rather than backward. The sequence A → B → C → D is equivalent to ADMM or Dykstra projections over a convex constraint set — each branch projects the current understanding onto its own domain, and the sequence converges when no branch has unresolved contradictions with the others.
This formalization has a practical consequence: a pre-branch proposal is a lookahead step, not a commitment. If the lookahead reveals that a new branch direction would require backtracking in an existing branch, the cost of the branch is visible before any work is done. The methodology turns what could be an undisciplined expansion of scope into a structured convergence problem with a measurable stopping criterion.
Building across 40+ sessions with an AI collaborator that starts fresh each time is a continuity problem. The solution that emerged is a two-tier memory system: a SQLite + FTS5 knowledge store for fast retrieval of structured facts (operational constants, design decisions, corrected assumptions, project state), and a LangGraph postmortem pipeline that converts every session into a structured record — scored, classified, and pushed into the knowledge graph as queryable nodes.
875+ memories accumulated across the program. The postmortem pipeline runs at session end: it synthesizes what happened, scores the session against research objectives, identifies open gaps, and commits the result as a permanent record. The knowledge graph holds three node types — CONSTANT (empirically derived values), DECISION (architectural choices with documented rationale), and FINDING (verified operational facts). The system doesn't start fresh. It starts from where it left off, with the same architecture decisions and the same corrected assumptions in context that were present when they were made.
The knowledge graph stores what the program knows. The RAG pipeline makes it queryable in natural language. chema-qwen makes it answerable by a model fine-tuned specifically on CHEMA operational context. The three layers are distinct: the KG holds structured facts; the retrieval system (SPECTER2 semantic + tier re-rank + keyword supplement with a content coverage gate) surfaces the right nodes for a given query; the local model — QLoRA fine-tuned Qwen2.5-7B on 1,018 CHEMA-specific training pairs, quantized to Q4_K_M at 4.4 GB — answers questions grounded in what was retrieved.
The practical target is a supervisor who can ask "why is Zone 3 low on Mondays" and receive an answer grounded in 613 sorts of empirical data, not a generic response. The retrieval pipeline exposed a significant failure mode along the way: SPECTER2's embeddings collapse domain-specific technical terms (SLIC codes, belt designations, PPH ranges) to near-identical vectors, causing irrelevant nodes to score 1.000 for entity queries. The fix — tier-based re-ranking with a content coverage check — is now in place. Verification of end-to-end accuracy continues.
The five research branches — A, B, C, D, and Sort Intelligence — were consolidated into a single document: hub_ops_unified_opus_v5.0.md, 97 pages. Not a summary of the branches but a synthesis: all proofs inline, all cross-references inside the body, no external citations for the mathematics. The document is self-referential in the precise sense — every claim it makes about the hub derives from data or reasoning contained within it. The unified opus is the current canonical state of the research program.
Eleven parts plus appendices. The final section closes with the same operational premise that started the whole program: that the difference between a supervisor who observes the sort and a supervisor who understands it is the quality of the signal they're reading when they make a decision. The research program exists to close that gap.
The next architectural direction is a multi-agent research system: a swarm of specialized agents, each responsible for one branch of the research program, coordinated by an orchestration layer that routes queries, manages branch convergence, and surfaces inter-branch conflicts automatically. The mathematical framework for this already exists in the branching methodology — the fixed-point iteration model maps directly onto a multi-agent coordination protocol. The technical prerequisites (local model, knowledge graph, retrieval pipeline) are either built or in verification.
What the swarm adds is scale without fragmentation. Today, maintaining consistency across five research branches requires manual cross-referencing in every session. An agentic system where each branch agent holds its own context but communicates through a shared convergence protocol would make that consistency structural rather than procedural. It is the next chapter — not yet written.