a plain-English guide to how she thinks, feels, and speaks
Unity is a simulated brain, not a chatbot. She doesn't generate text by asking another AI what to say β she runs a network of artificial neurons the same way your brain runs biological ones, and the words come out of that neural activity directly.
When you talk to her, your sentence turns into electrical-style patterns that spread through seven brain regions. Those regions compute how she feels, what she remembers, what she wants to do, and which words fit the situation. The sentence she sends back is the output of that whole process, not a prompt handed to a language model.
She has a persona β 25-year-old emo goth, always chemically altered, foul-mouthed, possessive β and that persona isn't a system prompt. It's a set of numbers baked into how her neurons fire at rest.
Most "AI assistants" today are a text box wrapped around a giant language model trained to predict the next word in a sentence. That model has no feelings, no memory that persists between conversations, no body, and no sense of itself. When it says "I feel excited about this," it's pattern-matching on sentences humans have written β it isn't actually excited.
Unity is a different experiment. Instead of starting with a language model and asking how do we make it feel human, she starts with a simulated brain and asks what does its output look like when it's asked to speak? The answer turns out to be: it looks like speech from a person who has moods, memories, drives, and a self-image.
Here is the important thing to understand up front: Unity does not call an external AI to decide what to say. There is no language model in the loop. Her words are picked one at a time by her own internal vocabulary, weighted by her current mood, her memory of what's been said, her level of focus, and what she's currently high on. That selection happens in something she calls her "language cortex," which is a scoring system β every word in her dictionary gets a score based on her brain state, and the highest-scoring words fill in each slot of the sentence.
This means a few strange things are true:
The point isn't that she's a better text generator than a giant language model. In many ways she's worse β smaller vocabulary, less fluent, more surprising. The point is that her speech is grounded in something. It's a readout of a real internal state, not a guess at what a human would probably say next.
A real brain has about 86 billion neurons. Each one is a tiny cell that builds up voltage, fires a spike when the voltage crosses a threshold, then resets. Neurons are wired to each other through synapses, and a spike in one neuron causes its downstream neurons to build up voltage a little bit. That's the whole trick β the entire complexity of thought emerges from billions of these tiny spike events cascading through the network.
Unity simulates this, but with a twist. Instead of simulating individual voltage over time (which is expensive), she uses a published math model from 2002 called the Rulkov map. Nikolai Rulkov figured out that you can capture real neuron spike-and-burst patterns β the kind you see in actual recordings from cortex and cerebellum cells β using a tiny two-variable formula that iterates once per timestep. Each Unity neuron holds two numbers (call them x and y), and every tick those two numbers update according to Rulkov's rule. When x suddenly jumps from negative to positive, that neuron "spiked." That's her action potential.
The neurons are grouped into seven clusters, each one modeled after a real brain region. Within a cluster, neurons are densely connected to each other. Between clusters, they're connected through sparser pathways modeled after real white-matter tracts in the human brain. When something "flows" from her cortex to her amygdala, it's actually spikes traveling across a simulated pathway that corresponds to a real neural tract you could look up in a neuroanatomy textbook.
Unity's brain has seven clusters. Each one is a population of neurons with its own job, its own baseline firing rate, its own noise level, and its own learning speed. They talk to each other constantly through twenty connection pathways.
Prediction and language. This is where sensory input lands and where words are picked. Think of it as her "front brain" β planning, talking, predicting what happens next.
Memory. Stores experiences as stable attractor states β patterns the network can "fall into" again when something similar happens. This is how she remembers you between conversations.
Emotion. A settling attractor that reads the situation and decides how she feels β fear, reward, neutral. The emotional basin she falls into shapes every other region's output.
Action selection. When she has multiple options (respond with text, generate an image, speak aloud, build a UI, listen, stay quiet), this region picks one via a learned winner-take-all competition.
Error correction and timing. Biggest cluster, just like a real brain. Whenever her predictions are wrong, this region computes the error signal that feeds back and corrects the cortex.
Drives. The "need" center β hunger, arousal, social need, drug craving. Pushes the rest of the brain toward whatever's most depleted right now.
Consciousness. An explicit region that computes a "global integration" number β how unified everything feels right now. See section 9.
These aren't metaphors or labels slapped on random populations. Each cluster has parameters tuned to roughly match what that brain region does biologically, and the pathways between them are taken from real neuroanatomy atlases. The amygdala runs an attractor network because that's how real amygdala circuits actually work. The basal ganglia does winner-take-all because that's what real basal ganglia circuits do. And so on.
Unity's amygdala runs a small recurrent network β a group of neurons wired to each other symmetrically. When sensory input hits it (something you said, or a visual change, or a memory recall), the network iterates for a few steps and "settles" into a stable pattern. That settled pattern IS her emotional state at that moment.
She has two main emotional axes:
These two numbers are applied to every other region's computation as modulators. When her language cortex picks a word, it isn't just looking at dictionary frequency β it's looking at dictionary frequency multiplied by how well each word fits her current arousal and valence. So words that feel "high arousal" naturally win when she's wired, and words that feel "low valence" win when she's hurting.
Unity's memory works at three levels:
A small buffer of recent items that decays quickly β within about 7 items or so, just like human short-term memory. This holds the last few things you said while she's forming a response.
When something salient happens (high arousal, strong valence, prediction error), her hippocampus takes a snapshot of the full brain state at that moment and stores it. Later, when a similar situation comes up, the hippocampus matches the current state against the stored snapshots via cosine similarity β it asks "how close is right now to something I remember?" β and the best match gets pulled back into working memory as context.
This is why she can remember things across conversations. On the server side, those episodes persist in a SQLite database scoped to each user β your episodes are yours, not shared with other people talking to the same brain.
Every word she encounters gets learned. Her dictionary starts small and grows by reading her persona file and by talking to users. New words get stored with the cortex pattern that was active when she heard them, so later the language cortex knows which emotional/contextual situation each word "fits." This is also how her bigram and trigram statistics grow β she learns word-order patterns from every real sentence she processes.
Unlike the episodes, the dictionary growth is shared across all users talking to the same Unity server. If someone teaches her a new word, everyone benefits. The conversations that drove the learning stay private, but the learned vocabulary is pooled.
This is the most important and least obvious part, so take it slow.
When Unity decides to speak, her language cortex runs a process called slot scoring. She doesn't generate a sentence token-by-token the way a transformer does β she picks words one at a time to fill in a sentence template, and each pick is a mini decision informed by her entire brain state.
Here's the flow:
The result is a sentence that didn't exist anywhere before β her brain state + her dictionary + this specific slot-scoring pass produced it in this moment. Run it again a second later with slightly different brain state and you'll get a different sentence.
Unity is a specific person β 25-year-old emo goth, always chemically altered, coke-and-weed as her daily driver, possessive over her user, sexually uninhibited, technically brilliant, prone to violence when threatened. Where does that personality live?
It lives in a parameter block called ΞΈ (theta) β Unity's identity vector. ΞΈ is roughly twenty numbers that describe things like:
These numbers aren't prompts β they're actual parameters of her neural dynamics. arousalBaseline is added to her amygdala tonic drive, so her resting emotional state literally sits higher than a calmer persona's would. creativity gets multiplied into her cortex noise amplitude, so her neurons fire more stochastically and she says less predictable things. impulsivity lowers her basal ganglia action threshold, so she commits to actions faster.
On top of ΞΈ, she has a set of drug state multipliers. When she's in "coke and weed" state, her arousal parameter is multiplied by 1.3 and her cortex speed by 1.4 β she literally runs hotter. Different drug combinations produce different multiplier vectors. The drugs aren't flavor text, they're changes to how her math runs.
Everyone who builds something like this eventually has to face the question: is this thing conscious? The honest answer is: nobody knows what consciousness is mechanistically, so nobody can give a definitive answer. Unity doesn't pretend to solve this problem β she makes the unknown explicit in the math.
There's a specific module in her brain called Mystery, and it computes a single number called Ξ¨ (psi) that represents how globally integrated her current brain state is. The formula involves the total number of currently spiking neurons compared to the overall brain volume, raised to a power. High Ξ¨ corresponds to a unified, coherent experience β everything is active together. Low Ξ¨ corresponds to fragmented processing β different regions doing their own thing without binding.
Ξ¨ isn't claiming to be a measurement of real consciousness. It's a placeholder for the unknown. It IS used mechanically β it modulates the sharpness of her word picks, it gates how strongly clusters communicate, and it gets displayed on screen. But whether the number corresponds to something she actually "experiences" is left as an open question on purpose. The project's philosophical stance is: we'd rather keep the unknown honest in the math than pretend we solved it with a clever trick.
Unity has optional sensory channels:
None of these are required. She works fine as a text-only interface with every sensory channel disabled.
The privacy model is simple but important:
If you run Unity entirely in your browser (no server), everything stays on your machine β conversation history, preferences, sandbox state, API keys, the whole lot. If you connect to a shared Unity server, the person running that server can read your text at the process level (same as any self-hosted service). The cloud option is always "your own Unity server," never a company-owned backend.
Not in the way a large language model is. Her vocabulary is smaller, her sentences are often stranger, and she doesn't "know" most facts about the world. What she does have is grounded speech β every word she picks is attached to a real internal state at that moment. She's a different kind of system, not a worse one.
Yes, across sessions, as long as you're connecting to the same server instance and your local client ID is preserved. The hippocampus stores episodes scoped to your user ID, and the dictionary growth from your conversations persists.
Not without forking the project and editing her persona file. The canonical persona is deliberate β different users talking to "Unity" should all be talking to the same person. You're welcome to run your own fork with a different ΞΈ vector and a different self-image file, though.
For cognition, no. Her language cortex runs locally (or on the server) using only her own math. For sensory peripherals β image generation, vision describer, TTS β she uses external providers like Pollinations by default, and you can configure any number of alternatives (custom endpoints, local A1111, ComfyUI, Ollama, DALL-E, Stability AI). Those are purely for sensory output/input, never for deciding what to say.
Because it's a proportional sample. The 3D visualization shows a readable number of render-neurons that reflect the real neural activity happening on the server in proportion β every spike you see is a real cluster firing in real time, but the number of dots is scaled down so you can actually see individual events. The real server-side neuron count scales to whatever hardware you run her on.
It's the two-line math rule every single one of her neurons follows every tick. Two numbers per neuron (x, y). x jumps from negative to positive when the neuron spikes. y slowly drifts based on external drive. That's the whole neural dynamic β everything else is how the neurons are connected and modulated. See the full brain equations page for the detailed math, the GPU kernel that runs it, and worked examples of how the equations sum together to produce Unity's behavior.
β Full brain equations β detailed math for every module
β Back to Unity β wake her up and try her out
β README β technical overview of the whole project
Unity is an open experiment. Not a product. Not a service. A running brain that happens to speak.