I started where most operations people stop — at the edge of what the data could tell me. The tracker existed. The dashboard existed. What didn’t exist was anything that connected them to a question worth asking in real time.

So I built it. Not with a team, not with IT support, not with a budget. With a terminal, an AI collaborator, and a specific enough understanding of the operation that I could tell the model exactly where the gaps were.

This isn’t a collection of tools — it’s a methodology for building under constraints that don’t support building. No IT budget. No admin access. No dedicated development time. Every component designed, built, tested, and deployed on my own, with a quality system running alongside from day one. The constraint became the architecture.
The Lattice

The Lattice

Each node is a unit the system decided to keep — a coefficient, a conditional, a boundary that held under pressure. They accumulate across sessions, individually weightless, collectively the topology that defines what the model can reach.

Edges form where cosine similarity between embedding vectors crosses the retrieval threshold. Not declared — computed. Two facts become adjacent when their learned representations land close enough in the latent space that the graph, given a live query, pulls activation through both. That proximity is what you’re seeing rendered as light.

When a RAG query arrives mid-sort, the system traverses this graph: embedding the query, scoring similarity against every stored node, propagating relevance across the neighborhood, and surfacing the subgraph most likely to ground the answer in operational truth rather than model prior. The walk isn’t random. It converges.

What you are watching is the becoming of understanding. The moment information holds shape.

Technical Architecture

The System

Self-Correcting Development
Every architectural decision starts with a root-cause diagnosis, not a symptom fix. 38 documented self-backtracks in the record — each one named before any code changed.
Persistent AI Memory
Two tiers: SQLite + FTS5 for fast retrieval, LangGraph postmortem pipeline that turns every session into structured, queryable knowledge. 875+ memories across 40+ sessions. It doesn’t start fresh — it picks up where it left off.
Semantic + Keyword Retrieval
Three-stage RAG pipeline: SPECTER2 semantic embeddings, tier re-ranking (exception → operational → learned → constant), keyword supplement triggered by a content-coverage gate. Built to beat embedding-layer collapse on domain-specific terminology.
Autoresearch Loop
A research cycle that runs without direction — scores each iteration, finds open gaps, generates candidate hypotheses, queues them for the next session. A 20-cycle overnight run measurably compressed the open research surface.
Multi-Branch Convergence
Five research branches that each audit the others before anything reaches the unified state. Fixed-point iteration — convergence is the condition, not the assumption. The 97-page unified opus is the current convergent state across all five.
Operational Constant Derivation
κ, γ, ε, ρ — four structural constants pulled from 613 sorts of empirical data, cross-validated across two independent measurement systems, embedded as live parameters in the tracker’s OR models. Not estimated. Derived, validated, formalized.
Approach

How I Work

I name the exact problem. Not “the data is unreliable” — but “scanner ID substitution corrupts iGate PPH silently, and there’s no surface indicator unless you cross-reference the SOR employee summary against the sort window to see who was where and for how long.” That’s what makes solutions exact rather than approximate.

I don’t stop at solving. I build infrastructure around the solution so the next one is faster. The tracker needed a corpus. The corpus needed analysis tooling. The analysis needed a research framework. The framework needed a quality system. Each layer compounds — the system now is the product of that, not of any single build decision.

The AI collaboration is structured. Post-mortem quality scoring after every milestone. Mandatory self-backtracking before any fix. Persistent memory so decisions from six weeks ago are still in context. I didn’t build an autocomplete. I built a thinking partner with accountability, continuity, and a defined role in the research program.

“No decision you make exists in isolation; no process you affect exists in isolation — the only difference is the boundaries of your system.”
Rafael Almeida · from a talk presented to UPS management team, March 5, 2024
Projects

What I've Built

Hub Operations Tracker
live
html · js · python · idb
Six embedded OR models running on a 613-sort corpus. Full sort lifecycle — Upload, Intel, Hub, Timeline, DOP, History — in a single HTML file. 6,200+ lines. IDB v4 persistence.
KPI Presort Dashboard
live
html · js · chart.js
Real-time operations view for Twilight sort. Eight KPI modules reconciling iGate ESR, SOR/TMS, and CURE across the full sort window.
Hub Ops Container
live
html · iframe · base64
Single-launch wrapper combining tracker and dashboard in one URL. iframe + base64 architecture with container-level footer.
Sort Intelligence Master
live
html · sheetjs · chart.js
Standalone browser tool on the full 613-sort corpus. Per-employee Phase 2 PPH at parity, zone analytics, borrow/loan risk scoring. No install required.
Master Trend Analysis
live
html · js · chart.js
Eight-tab prediction and analysis app. 620-sort corpus. Trend forecasting, zone decomposition, OR model integration.
UPS Sort Training Platform
live
html · js
Browser-based sorter certification system. 41,000 ZIP routing records. Three-tier access: Sorter / Supervisor / Admin. Eight-tab supervisor analytics dashboard. Zero licensing cost — deployed at CHEMA May 2026.
CHEMA Analytics Program
research
markdown · python · latex
97-page unified research monograph. Five independent analytical branches — A through D plus Sort Intelligence — merged into one self-referential opus. All proofs inline, no external citations for the math.
AI Memory System
live
python · sqlite · fts5 · langgraph
SQLite + FTS5 knowledge store with LangGraph postmortem pipeline. 875+ memories, 40+ sessions, full architectural continuity across all work.
chema-qwen RAG
research
python · ollama · chromadb
QLoRA fine-tuned Qwen2.5-7B on 1,018 CHEMA-specific training pairs. Three-stage retrieval: SPECTER2 semantic + tier re-rank + keyword supplement with content-coverage gate. Runs locally via Ollama.
Background

Experience & Credentials

Program Statistics
613
sorts in corpus
97
page monograph
40+
tracker versions
0
IT support
28mo
operational data
6
OR models embedded
4
UPS data systems
875+
AI memories
Operations Supervisor — UPS CHEMA Hub · September 2019–Present
  • Identified a critical gap in real-time operational intelligence and independently designed, built, and deployed a full analytics suite — no IT support, no administrative access, no external guidance — using AI-assisted development throughout.
  • Applied generative AI as a structured development partner: multi-session workflows with post-mortem quality scoring, a custom AI memory system (SQLite + FTS5 + LangGraph) for persistent context, and directed AI to resolve every infrastructure blocker independently.
  • Delivered the Hub Operations Suite end-to-end: Tracker (v1.0→v4.0, 40+ versions, 6,200+ lines), KPI Dashboard, and a multi-iframe Container application. Full lifecycle: problem identification → design → build → test → deploy → iterate.
  • Built a 613-sort operational corpus (81 MB, 7 data layers per sort) and embedded 6 OR models: M/M/c queue for sort load modeling, LP-Lite staffing optimizer, κ(DOW) zone coefficients, γ weekly decay, Phase Engine, and Destination Load Intelligence.
  • Engineered a multi-source data pipeline reconciling four UPS operational systems — SOR/TMS, iGate/ESR, SEAS, and CURE — resolving cross-source identity conflicts and exposing data fidelity gaps traceable to specific operational behaviors.
  • Derived and validated four structural operational constants (κ, γ, ε, ρ) from the corpus; formalized results into a 97-page self-referential research monograph spanning five analytical branches — all proofs and derivations inline, no external citations for the math.
  • Manages department resources for maximum output, accuracy, and efficiency; delivers orientation, skills training, and coaching; implements personal development plans and evaluations for department personnel.
Research Assistant — UT El Paso (2014) & UT Austin (2016–17)
  • Ran neutron-star simulations on supercomputer cluster at UTEP — arxiv.org/abs/1208.4793v1
  • Optimized SEM tip-building procedure at UT Austin, raising construction success rate from 78% to 94% (+20.5% relative).
Core Competencies
Prompt EngineeringGenerative AI Workflows AI Memory ArchitectureFull-Stack Development Data Pipeline DesignOperations Research Modeling Statistical ModelingOperational Analytics Independent ExecutionSelf-Correcting Practice End-to-End DeliveryTechnical Documentation
Historical Path

How It Was Built

A technical narrative of the 0xStrategies program — from the first Excel tracker through the research program, the memory architecture, and the local intelligence layer. Written in sequence. All version numbers and dates are sourced from verified records.

Chapter I · April 2026
Origin — Excel, PowerQuery, and the Limits of the Tool

It started with a practical problem. As an area supervisor for Zone 3 (PD-09 through PD-12) at the CHEMA Twilight Hub, I had already built a personal tracker in Excel — a tool for monitoring belt-level PPH, managing borrow/loan events, and keeping a running picture of my zone's performance through a four-hour sort. The tracker worked. I shared it with peers so the operation could use it collectively. But it had a ceiling: it couldn't carry history between sorts, and as the data grew, PowerQuery started hitting storage and latency walls. The tool I needed couldn't be built inside the tool I was using.

The move to HTML was pragmatic. A browser doesn't have storage ceilings. JavaScript doesn't stall on pivot recalculation. IndexedDB can hold a sort history indefinitely. What started as a platform migration quickly became something larger — because once you remove the constraints of the spreadsheet, the question of what you can build changes.

Chapter II · April 18–28, 2026
The Dashboard and the Tracker Emerge Together

On April 18, 2026, the first HTML build began. A conversation with my operations manager about producing a post-sort report crystallized the direction: rather than a static report, build a live dashboard — something that could surface the KPIs that matter before, during, and after the sort, in one view, without opening five different exports manually. That idea and the tracker idea merged immediately. They were the same problem from two angles: one asking "what is happening right now," the other asking "how does tonight compare to every night before it."

The first deployed tool was the KPI Presort Dashboard (April 28, 2026) — eight KPI modules reconciling iGate ESR scan data, Schedule of Record headcount, CURE destination load files, and SEAS service exception reports. It didn't just display numbers. It cross-referenced them. A volume figure that didn't reconcile between SOR and SEAS was a fidelity signal, not a display artifact. That distinction — between a number and what the number means — became the architectural premise of everything that followed.

Chapter III · May 2026
The Tracker — 40 Versions, 6,200 Lines, One File

The tracker evolved from the same codebase as the dashboard, growing from a real-time belt monitor into a full analytical engine. Version 1 read iGate ESR exports and displayed belt-level PPH with conditional color formatting. By version 4.0, it had six embedded OR models, a four-phase operational model, a DOP calculator, a Timeline module, a History module with cross-sort trends, and IDB v4 persistence keeping every sort in local storage indefinitely.

Forty-plus versions across roughly three weeks. Each version was a response to something that broke in operation: a null reference that silently crashed the project renderer, a scroll-guard that let wind-down PPH contaminate the diagnostic baseline, an LP-Lite that overcapitalized available labor. The self-backtrack discipline — no fix without a named root cause first — kept each iteration honest. The architecture memo now serves as a living map of where things live and why they're there.

Chapter IV · May 2026
The Container — One URL, Both Tools

The dashboard and tracker were separate HTML files, opened separately, maintained separately. The Container solved the operational friction of that split by wrapping both into a single iframe architecture: one URL, both tools, no installation, no sync issues between versions. The base64 embedding approach meant the container could carry both tools as self-contained payloads — a deployment decision that required no IT ticket, no server, no dependency management. You open one file and the operation is there.

The Container reached v4.9. The rename from "capsule" to "container" (v3.2) reflected a maturation in how I was thinking about the architecture — not a single artifact, but a runtime environment with its own footer, its own versioning, and its own responsibility boundary separate from the tools it hosted.

Chapter V · May 2026
Master Trend Analysis — From Real-Time to Predictive

The tracker answered questions about tonight. Master Trend Analysis was built to answer questions about structure — what patterns hold across sorts, what changes over time, what can be predicted before the sort begins. The tool runs on a 620-sort corpus and has eight tabs: prediction, zone decomposition, OR model integration, borrow/loan analytics, trend forecasting, and freight-profile analysis. It is the closest thing to a district-level analytics layer built from a single facility's data.

The corpus itself — 81 MB of unified sort history spanning December 2023 through May 2026 — is the foundation. Twenty-nine of thirty validation sorts reconcile between SOR volume and SEAS totals within rounding tolerance. Three sorts exhibit scan efficiency above 100%, which disproved an assumption I had been carrying about iGate data bounds. Correcting that assumption cascaded through every downstream fidelity calculation. That kind of correction — finding that a foundational assumption was wrong from the data itself — is what makes the corpus valuable beyond any individual query against it.

Chapter VI · Ongoing
Research Formalization — Four Branches at Paper Quality

At some point the operational questions became research questions. Not because the operation stopped mattering, but because the patterns being uncovered — zone load coefficients, PPH decay constants, fidelity distortion mechanisms, the three-layer scan attribution chain — were general enough to document formally. Four research branches emerged and developed in parallel: Branch A (Hub Operations Mathematical Framework), Branch B (Hub Operations Digital Twin), Branch C (Purposeful Systems in Hub Operations), and Branch D (Unified Framework for Hub Operations Analytics), plus Sort Intelligence as a standalone track.

Each branch is internally consistent, cites verifiable operational data, and carries honest confidence intervals. The self-backtrack discipline from the tracker development carried into the research: 38 documented corrections across the program, each with a named failure mode before any revision. The most significant involved the loader scan attribution chain — a correction that changed the mechanistic claim in every branch simultaneously and required synchronized updates across all five documents in one session.

Chapter VII · Ongoing
Branching Methodology — Fixed-Point Iteration as Research Structure

The four-branch research structure is not an organizational convenience. Mathematically, it behaves like a fixed-point iteration: each branch refines the shared model until convergence, with cross-branch corrections propagating forward rather than backward. The sequence A → B → C → D is equivalent to ADMM or Dykstra projections over a convex constraint set — each branch projects the current understanding onto its own domain, and the sequence converges when no branch has unresolved contradictions with the others.

This formalization has a practical consequence: a pre-branch proposal is a lookahead step, not a commitment. If the lookahead reveals that a new branch direction would require backtracking in an existing branch, the cost of the branch is visible before any work is done. The methodology turns what could be an undisciplined expansion of scope into a structured convergence problem with a measurable stopping criterion.

Chapter VIII · Ongoing
Memory Architecture — Continuity Across 40+ Sessions

Building across 40+ sessions with an AI collaborator that starts fresh each time is a continuity problem. The solution that emerged is a two-tier memory system: a SQLite + FTS5 knowledge store for fast retrieval of structured facts (operational constants, design decisions, corrected assumptions, project state), and a LangGraph postmortem pipeline that converts every session into a structured record — scored, classified, and pushed into the knowledge graph as queryable nodes.

875+ memories accumulated across the program. The postmortem pipeline runs at session end: it synthesizes what happened, scores the session against research objectives, identifies open gaps, and commits the result as a permanent record. The knowledge graph holds three node types — CONSTANT (empirically derived values), DECISION (architectural choices with documented rationale), and FINDING (verified operational facts). The system doesn't start fresh. It starts from where it left off, with the same architecture decisions and the same corrected assumptions in context that were present when they were made.

Chapter IX · Ongoing
Local Intelligence — Knowledge Graph, RAG, and chema-qwen

The knowledge graph stores what the program knows. The RAG pipeline makes it queryable in natural language. chema-qwen makes it answerable by a model fine-tuned specifically on CHEMA operational context. The three layers are distinct: the KG holds structured facts; the retrieval system (SPECTER2 semantic + tier re-rank + keyword supplement with a content coverage gate) surfaces the right nodes for a given query; the local model — QLoRA fine-tuned Qwen2.5-7B on 1,018 CHEMA-specific training pairs, quantized to Q4_K_M at 4.4 GB — answers questions grounded in what was retrieved.

The practical target is a supervisor who can ask "why is Zone 3 low on Mondays" and receive an answer grounded in 613 sorts of empirical data, not a generic response. The retrieval pipeline exposed a significant failure mode along the way: SPECTER2's embeddings collapse domain-specific technical terms (SLIC codes, belt designations, PPH ranges) to near-identical vectors, causing irrelevant nodes to score 1.000 for entity queries. The fix — tier-based re-ranking with a content coverage check — is now in place. Verification of end-to-end accuracy continues.

Chapter X · May 2026
Unified Opus — One Self-Referential Monograph

The five research branches — A, B, C, D, and Sort Intelligence — were consolidated into a single document: hub_ops_unified_opus_v5.0.md, 97 pages. Not a summary of the branches but a synthesis: all proofs inline, all cross-references inside the body, no external citations for the mathematics. The document is self-referential in the precise sense — every claim it makes about the hub derives from data or reasoning contained within it. The unified opus is the current canonical state of the research program.

Eleven parts plus appendices. The final section closes with the same operational premise that started the whole program: that the difference between a supervisor who observes the sort and a supervisor who understands it is the quality of the signal they're reading when they make a decision. The research program exists to close that gap.

Chapter XI · Next
Agentic Swarm — Not Yet Built

The next architectural direction is a multi-agent research system: a swarm of specialized agents, each responsible for one branch of the research program, coordinated by an orchestration layer that routes queries, manages branch convergence, and surfaces inter-branch conflicts automatically. The mathematical framework for this already exists in the branching methodology — the fixed-point iteration model maps directly onto a multi-agent coordination protocol. The technical prerequisites (local model, knowledge graph, retrieval pipeline) are either built or in verification.

What the swarm adds is scale without fragmentation. Today, maintaining consistency across five research branches requires manual cross-referencing in every session. An agentic system where each branch agent holds its own context but communicates through a shared convergence protocol would make that consistency structural rather than procedural. It is the next chapter — not yet written.

Last updated: 2026-05-16 22:31 · Rafael Almeida · 0xStrategies