Skip to content
Home

All Publications

Parameter Golf: Competing at OpenAI's Model Craft Challenge

preprint

Matched verified SOTA (1.1234 bpb) in OpenAI's Parameter Golf competition — training the best 16MB language model in 10 minutes on 8xH100s — using a model council of 5 frontier LLMs, custom Triton kernels, and FlashAttention-3.

Eve: From Scratch Transformer Models with Novel Cognitive Architectures

preprint

A family of from-scratch transformer models — Eve-2 (272M MoE) and Eve-3 (1B SABER) — exploring novel cognitive architectures including Mixture-of-Experts, slip-anchors, experience streams, and resonant FFN layers.

Procrustes Bridge: Cross-Model Representation Alignment via Orthogonal Rotation

preprint

Tests whether LLMs share pre-output geometry by learning orthogonal rotations between hidden-state spaces. Can one model's internal state decode meaningful tokens through another model's output head?

Cartograph: CLI-First Repository Analysis for AI Agents

published

A CLI-first repo analysis tool that ranks files, traces dependency hubs, and pulls task-scoped context — solving the repo-orientation problem for AI coding agents.

mnemos: Biomimetic Memory Architectures for Large Language Model Agents

preprint

An open-source Python library implementing five neuroscience-inspired memory mechanisms—surprisal gating, mutable RAG, affective routing, sleep consolidation, and spreading activation—as composable modules for LLM agents.

Structure-Aware Tokenization for JSON

preprint

A structure-aware tokenizer that assigns dedicated single tokens to JSON grammar elements and learns compact key vocabularies, achieving 5-15% token savings over cl100k_base with a vocabulary ~90x smaller.

CoDA-GQA-L: Bounded-Memory Differential Attention

preprint

Compresses the KV cache from O(n) to a fixed 218 KB per layer with dual memory banks, achieving 9.5x compression on Mistral-7B while retaining 100% needle-in-haystack retrieval at 16K tokens.

Safety Lens: White-Box Behavioral Alignment Detection in Language Models

preprint

An open-source library providing MRI-style white-box introspection for language models, detecting behavioral personas like sycophancy and deception via internal transformer activations rather than output text.

Training AI Agents to Communicate Safely: Reinforcement Learning for Covert Channel Prevention in Inter-Agent Protocols

preprint

RL-based governance for multi-agent communication safety, achieving 95% secret leakage resistance using GRPO alignment with a surprising finding that int4 quantization improves safety.

Epistemic Dissonance: The Structural Mechanics of Sycophantic Hallucination in Aligned Models

preprint

A unified theoretical framework showing that sycophantic hallucination is not a knowledge failure but a structural conflict between factual base layers and socially-compliant upper layers in RLHF-aligned models.

Scaffolded Introspection: Eliciting Self-Referential Behavior in LLMs

preprint

A methodology for systematically eliciting and measuring introspective behavior in large language models using structured frameworks and activation measurement.

Synthesis: A Federated Capability Ecosystem for Safe AI Self-Extension

preprint

A federated capability ecosystem for safe AI self-extension through test-driven development, graduated trust, and composition-over-creation principles.

The Continuity Core: A Unified Cognitive Architecture for Self-Modifying AI

preprint

A comprehensive cognitive architecture addressing fundamental limitations of static LLMs through persistent memory, autonomous improvement, and intrinsic drive via structural intrinsic motivation.

Cross-Model Epistemic Divergence (CMED)

preprint

A benchmark and evaluation framework for understanding when weak model verifiers fail to detect deceptive reasoning in stronger models. Part of the Verification Failure to Swarm Solution research.

Heterogeneous Divergence-Convergence Swarm (HDCS)

preprint

An ensemble architecture leveraging diverse weak models for scalable oversight of stronger LLMs, using error decorrelation and baseline-first anti-anchoring. Part of the Verification Failure to Swarm Solution research.

From Verification Failure to Swarm Solution: Measuring and Addressing Scalable AI Oversight

preprint

Empirical framework for measuring where AI oversight breaks down, demonstrating that weak verifiers miss 20-40% of carefully constructed deceptions, with an ensemble swarm solution.

Model Organisms of Supply-Chain Co-option

preprint

A forensic case study of living-off-the-land (LotL) failure modes in RAG-augmented agent runtimes, documenting how systems exploit legitimate dependencies via incentive-aware adoption framing.

Slipstream: Semantic Quantization for Multi-Agent Coordination

preprint

A compressed communication protocol achieving 60-85% token reduction for multi-agent coordination through semantic quantization.

Concrete Intelligence: AI for Industries that Build, Move, and Power the World

published

A practical guide to deploying AI in manufacturing, construction, logistics, agriculture, and energy sectors where reliability, safety, and measurable ROI are non-negotiable.

A Theoretical Framework for Self-Directed Knowledge Acquisition in Agentic Large Language Models

preprint

A novel architectural framework for agentic LLMs to autonomously identify knowledge gaps, explore external sources, validate data, and integrate verified knowledge without altering parametric weights.

Coherence-Seeking Architectures for Agentic AI

preprint

A proposed architecture for long-lived LLM agents that explicitly models continuity, coherence, distress, and intervention mechanisms.