Research

We explore the structure of machine cognition. Understanding how neural networks think, reason, and represent knowledge.

Research Areas

Mechanistic Interpretability

Understanding neural networks by reverse-engineering their internal computations. Finding the circuits that implement specific behaviors.

Sparse Autoencoders

Training networks to decompose neural activations into interpretable features. Making the latent space legible.

Model Reasoning Inspection

MRI: A framework for understanding how language models reason. Tracing the path from input to output through interpretable steps.

Neural Topology

Studying the geometric and topological properties of neural network representations. The shape of thought in latent space.

The Three-Phase Model

Our approach to understanding neural network reasoning

Phase 1

Activation Capture

Recording neural activations across model layers during inference.

Phase 2

Feature Extraction

Decomposing activations into interpretable features using sparse autoencoders.

Phase 3

Reasoning Analysis

Tracing feature activations to understand the model's reasoning process.

Nomos Logos

Our AI research platform. A unified environment for interpretability research, model analysis, and reasoning inspection.

Currently in development. Join the waitlist for early access.

Request Access