The Vivarium AI Experimentation Platform
Nov 8, 2025
Posts in this series:
- The Vivarium AI Experimentation Platform
- Part 1: The Agents
- Part 2: The Languages
- Part 3: The Tools
- Part 4: The Infrastructure
- Part 5: The Roadmap
- Part 6: The Funding
Is Artificial Intelligence Real?
If you're like I was a few weeks ago, you may be tempted to believe that the reason the world is absurdly chasing LLMs on GPUs with hundreds of billions of dollars is because no one knows anything better about machine intelligence.
Like I was, you would be mistaken. And you can make a fair case for being excused for that. There's a lot of hype out there.
My goal in this series of posts is to present irrefutable evidence that an alternate future is possible, one where we collectively develop real machine intelligence that approaches human intelligence, using comparable amounts of energy, and accessible to all.
Sound like a fantasy, too good to be true? Well, hear me out first. After all, if my fantasy turns out to be impossible, you can always go back to LLMs and GPUs.
How Much Do We Know?
As it turns out, quite a lot going back many decades. A lot of the earlier "AI failures" had to do with an "over-application" of things we associated with "intelligence", like formal mathematical logic. With all due respect to Spock, it turns out that making sense out of the world isn't as simple as "reasoning".
Let's take a look at what is out there that's not LLMs. Here are what I consider three foundational books:
- Structural Information Theory by Emanuel Leeuwenberg. Do you want to learn about perception? Here's where to start.
- Machine Super Intelligence by Shane Legg. Did you know that we can define an optimal intelligent agent? Learn more here.
- An Introduction to Universal Artificial Intelligence by Marcus Hutter, Elliot Catt, David Quarel. Marcus Hutter was Shane Legg's advisor. There's a lot to learn from these two. (Also, they both worked at DeepMind.)
In addition to these books, here are projects pursuing artificial intelligence from perspectives very different than LLMs and genAI:
- Thousand Brains Project
- Open Brain Institute
- Hyper-dimensional Computing
- Free Energy Principle
- Bayesian Prediction / Active Inference
- Integrated Neuro-Symbolic Architecture
- Dynamic Field Theory
And here are three approaches to hardware that could be very promising:
- Extropic Thermodynamic Computing
- University of Texas at Dallas Neuromorphic Computer
- USC Viterbi School of Engineering and the School of Advanced Computing Integrated Spiking Artificial Neuron
What is Vivarium?
Vivarium is a comprehensive, integrated platform for the experimentation with and testing of arbitrary machine intelligence models. It is comprised of four core elements:
- AI agent definitions and models;
- Machine language and human-machine language interfaces;
- Tools;
- Infrastructure;
The totality of this system is called Vivarium.
The "AI agents" exist within a "world" and use "language interfaces" to "tools" and other agents. The worlds are described (structure and functionality) using languages capable of computation (i.e. not simply YAML or JSON, etc.).
The "infrastructure" is a combination of the computational resources (processor, memory, storage) to run the worlds and agents, and the mechanisms for organizing, structuring, and manipulating those computational resources.
Within the agents, an evolutionary mechanism can be introduced, again based on a language, that permits unguided exploration of solution spaces with respect to the structure or function of the agents.
The three other components are agnostic about the specific functioning of the "AI agent" component other than the agents' interface, which allows agents to interact with one another and with the world in which they are embedded and with the tools to which they have access.
What's in this Series of Posts
This series of posts introduces the core concepts in Vivarium:
- Agents: I will offer that we should be defining the terms we use to talk about "agents", including "agent" itself. There's a lot of prior art here, some of it linked above. Agents are the components of the system that contain "intelligence".
- Languages as interfaces: Unlike APIs, domain-specific languages as a medium for agents to collaborate with one another, as well as being the mechanism for how they use tools, presents a rich opportunity to encode capabilities and safety in the system.
- Tools are the means by which agents act: There's nothing new in this idea, but strictly enforcing these boundaries enables desirable system properties. Tools cannot act on their own. We can carefully define their capabilities, while still allowing for unbounded complexity in the agents.
- Infrastructure: The internet has been around for decades now and we are collectively still firmly stuck in the manufacturing era, where widgets are spit out of assembly lines into boxes into the shrink-wrapping machine onto trucks over highways into stores and onto shelves where we venture to review and sometimes purchase them. Don't let SaaS (software as a service) fool you; you're just renting that, it still comes in a box.
If that sounds interesting enough, you can certainly continue to Part 1.
The rest of this post will discuss the current AI bubble around LLMs and GPUs.
An Immensely Consequential Decision
Now is a good time to address the 800 lb gorilla in the room, or the 100s of billions of dollars gorilla, if you prefer.
"What about LLMs and genAI?" you may be tempted to ask.
Yes, indeed, what about them?
Let me start by telling you where I'm coming from here:
- I've spent more than 30 years writing complex software, like virtual machines and compilers.
- I understand a fair bit of math and some things about neural networks.
- I know a little about cognitive psychology, and comparative studies of intelligence, for example, with corvids and cephalopods.
- Over the past 18 months, I've used genAI intensely. Mostly ChatGPT, but also Gemini and Claude.
Maybe those are relevant or maybe not. I don't think Sam Altman has those and people listen to him. He didn't even build what he's selling, but maybe all we needed to invent AI was the right salesman to come along.
My own opinion based on the above is that genAI is a very wide cul-de-sac wherein some well-funded street racers are doing cookies in their custom built (at great expense) barely street-legal hot-rods and it's a blight on humanity. Not to put too sharp a point on it.
But I'm not going to be offended if you tell me I'm wrong, because I could be. But do you really think that one 10-page paper with an extremely humble title like "Attention is all you need" has cracked the nut of intelligence and we just need to spend a few hundred more billions of dollars to "scale" it?
Anyway, rather than taking my word for it, here are a couple links to other people's takes on it:
It's curious that every time someone invites some investor (in AI) to opine on whether there is an "AI bubble", somehow no one rings up Ed Zitron to ask if he's available to be on a panel.
Or you could ask ChatGPT, as I did, what are the fatal flows of the genAI/LLM approach and you might get something like the following:
- No Grounding in the Physical or Causal World (Semantic Grounding Problem)
- No Internal Causal Model of the World
- No Unified Self-Model or Persistent Memory
- No Embodiment or Situated Agency
- Lack of Compositional, Systematic Generalization
- Architectural Myopia: Sequence Prediction ≠ Thinking
- No True Intentionality or Semantic Reference
- Lack of Hierarchical, Cross-Modal World Models
- No Internal Curiosity, Goal Formation, or Motivation
- Epistemic Fragility and Lack of Meta-Cognition
The very idea that all there is to know about the world can be inferred by which word follows which other word should be, on its face, such a laughable notion, that people would indeed laugh when they hear it. But in fact, most don't, and a few throw tens of billions of dollars after it.
To Learn or Not to Learn
If we had persisted with the notion that the way to elevate oneself above the ground was to use a ladder because to everyone it appeared to usually work just fine, we never would have reached the moon.
If you're a critical observer, LLMs don't even appear to work much of the time, so continuing to pour billions of dollars into "scaling" them borders on an absurdity beyond the words to express it.
The idea that autonomous driving must "learn the city" in order to function obviously does not match our experience as human drivers, so even if it's the best we're capable of at this moment, it shouldn't be accepted as a reasonable approach to autonomous driving.
I'll go even further and plainly state that multi-trillion dollar market cap companies have no real utility or valuable contribution to the world economy. The goal should be putting AI in the service of every individual to enable them to live a happier, healthier, more fulfilling life alongside their friends and loved ones no matter where they are in the world. No company or group of companies should have a monopoly on that.
When one steps back just a little, it's very easy to see that the demo of ChatGPT was just too convincing. Like a young child does with their inanimate stuffed animal, we anthropomorphize something that gives the appearance of realness and imbue it with meaning and human attributes. The demonstrators were, of course, clever.
When the ChatGPT shimmers its "Thinking..." text after your prompt rather than telling the truth (crunching some small floats on these here GPUs), we eat it up. And then it tells us, "This is exactly the right question to ask and cuts right to the heart of blah blah blah" for the fiftieth time, and we keep on feeling quite proud of ourselves.
It is a little weird, though, that we do this. After all, intelligence isn't that rare on this earth. Just counting the humans, we're surrounded by about eight billion examples. And that's before we count the cats, dogs, birds, cows, horses, hummingbirds, and dragon flies. Surely, I've missed a few. LLMs can't do any of that. Not even dragon flies.
What a time to be alive.