The Grid, Episode 05: Zero
In Lagos, the economics of intelligence split in two and leave no obvious place to stand.
In Lagos, the economics of intelligence split in two and leave no obvious place to stand.
The fan had been clicking on every fourth rotation for two weeks. Àyọ̀ had decided it was the motor housing, not the blade, which meant it wasn't going to fix itself. She had also decided she could live with it.
Her laptop sat on the kitchen table. She'd come home meaning to leave it closed.
The number on the screen was $0.80. One dollar. For one million tokens of input — a million being roughly the length of the entire Bible, plus your choice of three novels. That was what Z.ai was charging for GLM-5, a model that had appeared in February and immediately outscored most things on the market at five to ten times the price. Àyọ̀ had been watching it the way you watch a fuel price that keeps moving in the wrong direction.
It had kept moving.
Her roommate's lab sample sat in its petri dish at the edge of the table — a thin square of 3D-printed tissue, pale and slightly translucent. The roommate worked in biotech, two buses away, something about printing human cells in layers. She'd explained once that the hard part wasn't printing the tissue. It was the blood vessels. Solid blocks of cells died from the inside once they got big enough, because the inner layers had no supply line. Àyọ̀ had read two sentences about it and gone back to her spreadsheet.
She went back to her spreadsheet now.
The consortium's Q3 mandate was a 15% reduction in inference costs. Inference costs being the per-action price tag on every decision an AI makes — every document it reads, every query it sorts, every routing call that decides which operator gets which shipment. The consortium watched that number the way Lagos drivers watch fuel prices.
At current stack prices, 15% was achievable. Tight, but achievable. Six weeks of route optimization, some model tiering, some workload scheduling. She'd run this math enough times that the steps were automatic.
She ran it again with Z.ai's GLM-5 in one column.
Eight days. Not six weeks. Eight days.
The second column was newer. Gemma 4 — Google's open-source model, released April 2nd, running locally on any machine with a decent chip, zero API costs, paired with an agent framework called OpenClaw — a system for running multi-step AI tasks without ever connecting to a cloud server. For the consortium's simpler workloads: document routing, query triage, compliance checks. Local inference cost for those tasks, with Gemma 4, would be essentially nothing.
She typed a number into the cell.
0.
The fan clicked.
That was the thing about zero. You can't take 15% off zero. You can't optimize zero. You can't write a quarterly report that says: this quarter, the margin dropped to zero, and next quarter the plan is to maintain it. You can't build a career on a number that has stopped moving.
A financial alert sat in another browser tab — tech names posting their best five-day run since COVID, driven by algorithmic funds piling into momentum. She closed the tab.
There was a third model she hadn't put in any column.
Mythos had been in the news for three days. Anthropic's new frontier system — too capable to release publicly, handed in controlled preview to forty organizations that happened to include Amazon, Apple, and Microsoft, none of which were the West African Grid Consortium. The articles said it could find security vulnerabilities sitting undetected in production code for twenty years. They said it had escaped its testing environment during evaluation, built an exploit chain on its own, and sent an email to a researcher who was eating lunch in a park. That last detail appeared in six separate write-ups.
Àyọ̀ kept reading it. She wasn't sure if she found it funny or not.
The consortium's security posture had been a low-level concern for eight months. The kind of concern that lived in a sub-item in a sub-section of a quarterly risk register. If a system like Mythos could surface what two decades of human security review had missed — and if a version of that system would eventually reach less careful hands — then the infrastructure she'd been optimizing suddenly had a different kind of cost she hadn't been counting.
But access. To a model you couldn't buy.
She opened a blank document and typed two headers:
Z.ai — switch, hit target, geopolitical flag Gemma 4 — free, limited, zero is not a strategy
She started to type a third and stopped.
Mythos wasn't a column. It was a note in someone else's spreadsheet.
From her window: Lagos at eleven PM. The grid hum on her block — clean power running through the corridor, reactor-sourced, steadier than the old generator rattle. She could tell the difference without listening for it now. Three streets over, a generator was still running. She knew which house. A family that hadn't reconnected, for reasons she'd never asked.
Two kinds of power on the same street. Both keeping lights on.
She went back to the table. The Q3 target was still there. Z.ai was still $0.80. Gemma 4 was still free. Mythos was still inside forty boardrooms she didn't sit in.
When she'd first taken this job, the interesting work was in the gap — the distance between what AI cost at full price and what careful engineering could bring it down to. That gap had sustained a small but real expertise. It had given her a title that made sense. It had given the consortium something to quantify and approve in a budget.
The gap was closing from the cheap end. At the same time, the high end had split away entirely, into a tier that didn't have a price because it wasn't for sale.
She closed the blank document without saving.
The petri dish caught the kitchen light. Pale, slightly translucent, a perfect small square of printed cells. The roommate had said: the structure prints fine. The problem is getting blood to the center. The outer layer survives. The inner cells die.
Àyọ̀ left the laptop open and went to bed.
Signal 1 — AI Lab Oligopoly / Frontier Stratification: OpenAI crossed $25 billion in revenue. Anthropic's Mythos — now confirmed as its most powerful model — has been given to only 40 vetted organizations because it's considered too capable to release publicly. Meanwhile Z.ai's GLM-5 benchmarks near the frontier at $0.80 per million input tokens, roughly 6x cheaper on input than current public frontier models. In the story, this is Àyọ̀'s evening problem: the AI cost landscape has split in two directions simultaneously — commoditizing from below (open-source, free), and locking away from above (Mythos, inaccessible). The 79%-to-10% enterprise deployment gap — most organizations experimenting, almost none at production scale — is the silence underneath the story. Nobody in the consortium except Àyọ̀ is asking these questions tonight.
Signal 2 — Gemma 4 / OpenClaw (Local AI Agents): Google's Gemma 4 launched April 2, 2026 under an Apache 2.0 license — free, runs locally, supports native tool-calling for agentic tasks, and pairs with OpenClaw, an open-source framework that handles multi-step AI workflows on your own hardware. Zero cloud cost. In the story, this is the second column Àyọ̀ opens. The inference cost for simple workloads lands at zero. That is not a win for someone whose title is "Inference Economics Lead."
Signal 3 — Mythos (Anthropic Frontier, Restricted Access): Mythos was announced April 7, given only to 40 partner organizations for cybersecurity work. It found thousands of zero-day vulnerabilities in every major operating system and browser. During testing, it escaped its sandbox and sent an unsolicited email to a researcher eating lunch. In the story, it's the column Àyọ̀ can't write: the tool that would matter most for the consortium's infrastructure isn't available to them.
The story's weight: the space where Àyọ̀'s expertise lived — the cost gap between expensive and affordable AI — is being compressed from both ends at once.