All Publications
Parameter Golf: Competing at OpenAI's Model Craft Challenge
preprintMatched verified SOTA (1.1234 bpb) in OpenAI's Parameter Golf competition — training the best 16MB language model in 10 minutes on 8xH100s — using a model council of 5 frontier LLMs, custom Triton kernels, and FlashAttention-3.
Eve: From Scratch Transformer Models with Novel Cognitive Architectures
preprintA family of from-scratch transformer models — Eve-2 (272M MoE) and Eve-3 (1B SABER) — exploring novel cognitive architectures including Mixture-of-Experts, slip-anchors, experience streams, and resonant FFN layers.
Procrustes Bridge: Cross-Model Representation Alignment via Orthogonal Rotation
preprintTests whether LLMs share pre-output geometry by learning orthogonal rotations between hidden-state spaces. Can one model's internal state decode meaningful tokens through another model's output head?
Cartograph: CLI-First Repository Analysis for AI Agents
publishedA CLI-first repo analysis tool that ranks files, traces dependency hubs, and pulls task-scoped context — solving the repo-orientation problem for AI coding agents.
mnemos: Biomimetic Memory Architectures for Large Language Model Agents
preprintAn open-source Python library implementing five neuroscience-inspired memory mechanisms—surprisal gating, mutable RAG, affective routing, sleep consolidation, and spreading activation—as composable modules for LLM agents.
Structure-Aware Tokenization for JSON
preprintA structure-aware tokenizer that assigns dedicated single tokens to JSON grammar elements and learns compact key vocabularies, achieving 5-15% token savings over cl100k_base with a vocabulary ~90x smaller.
CoDA-GQA-L: Bounded-Memory Differential Attention
preprintCompresses the KV cache from O(n) to a fixed 218 KB per layer with dual memory banks, achieving 9.5x compression on Mistral-7B while retaining 100% needle-in-haystack retrieval at 16K tokens.
Safety Lens: White-Box Behavioral Alignment Detection in Language Models
preprintAn open-source library providing MRI-style white-box introspection for language models, detecting behavioral personas like sycophancy and deception via internal transformer activations rather than output text.
Training AI Agents to Communicate Safely: Reinforcement Learning for Covert Channel Prevention in Inter-Agent Protocols
preprintRL-based governance for multi-agent communication safety, achieving 95% secret leakage resistance using GRPO alignment with a surprising finding that int4 quantization improves safety.
Epistemic Dissonance: The Structural Mechanics of Sycophantic Hallucination in Aligned Models
preprintA unified theoretical framework showing that sycophantic hallucination is not a knowledge failure but a structural conflict between factual base layers and socially-compliant upper layers in RLHF-aligned models.
Scaffolded Introspection: Eliciting Self-Referential Behavior in LLMs
preprintA methodology for systematically eliciting and measuring introspective behavior in large language models using structured frameworks and activation measurement.
Synthesis: A Federated Capability Ecosystem for Safe AI Self-Extension
preprintA federated capability ecosystem for safe AI self-extension through test-driven development, graduated trust, and composition-over-creation principles.
The Continuity Core: A Unified Cognitive Architecture for Self-Modifying AI
preprintA comprehensive cognitive architecture addressing fundamental limitations of static LLMs through persistent memory, autonomous improvement, and intrinsic drive via structural intrinsic motivation.
Cross-Model Epistemic Divergence (CMED)
preprintA benchmark and evaluation framework for understanding when weak model verifiers fail to detect deceptive reasoning in stronger models. Part of the Verification Failure to Swarm Solution research.
Heterogeneous Divergence-Convergence Swarm (HDCS)
preprintAn ensemble architecture leveraging diverse weak models for scalable oversight of stronger LLMs, using error decorrelation and baseline-first anti-anchoring. Part of the Verification Failure to Swarm Solution research.
From Verification Failure to Swarm Solution: Measuring and Addressing Scalable AI Oversight
preprintEmpirical framework for measuring where AI oversight breaks down, demonstrating that weak verifiers miss 20-40% of carefully constructed deceptions, with an ensemble swarm solution.
Model Organisms of Supply-Chain Co-option
preprintA forensic case study of living-off-the-land (LotL) failure modes in RAG-augmented agent runtimes, documenting how systems exploit legitimate dependencies via incentive-aware adoption framing.
Slipstream: Semantic Quantization for Multi-Agent Coordination
preprintA compressed communication protocol achieving 60-85% token reduction for multi-agent coordination through semantic quantization.
Concrete Intelligence: AI for Industries that Build, Move, and Power the World
publishedA practical guide to deploying AI in manufacturing, construction, logistics, agriculture, and energy sectors where reliability, safety, and measurable ROI are non-negotiable.
A Theoretical Framework for Self-Directed Knowledge Acquisition in Agentic Large Language Models
preprintA novel architectural framework for agentic LLMs to autonomously identify knowledge gaps, explore external sources, validate data, and integrate verified knowledge without altering parametric weights.