The Cognitive Architecture
An exploration of the internal mechanisms driving the Cyril Experiment.
Cyril is not just a language model. He is a cognitive architecture designed to evolve, remember, and understand you over time.
Unlike standard LLMs that predict the next token, Cyril operates in abstract latent spaces, learning to predict future states of the conversation and the relationship. This exposition breaks down the key components of his mind.
Hierarchical Latent Space
Thoughts, Sessions, and Life arcs represented as vectors.
Cyril processes information at multiple timescales simultaneously. Each timescale is represented by a distinct latent vector ($z$).
- z_thought (Fast)Immediate cognitive processing. Changes with every turn of the conversation. Represents the current train of thought.
- z_session (Medium)The context of the current interaction. Maintains the mood, topic, and immediate goals of the session.
- z_life (Slow)Long-term memory and personality evolution. Slowly adapts over days and weeks as he learns more about you.
"Hey, I really love hiking"
Predictive Learning (JEPA)
Learning by predicting the future in latent space.
The Joint Embedding Predictive Architecture (JEPA) allows Cyril to learn from interactions without needing human labels. He predicts the representation of future states based on the current context.
The model minimizes the distance between its prediction of the future state and the actual future state. Crucially, it predicts in abstract space (representation), not pixel or token space.
Variables
Interactive Loss Playground
Click on embeddings to toggle between context and target. The JEPA predictor learns to predict target embeddings from context using stop-gradient. Watch the loss change as you adjust targets.