Skip to content

Plutonic Rainbows

Cromptons of Ramsgate

The cabinet on the end of the row at Skegness was built before I was born. A two-penny coin slot, a sloped tray of copper, a hydraulic shelf shoving a tide of coins toward an edge that never quite spilled. The wood-effect side panel was patched with masking tape where someone had bashed it, probably more than once, and the back of the machine still ran on what looked like the original transformer. A boy in front of me dropped his last coin in. The shelf swept forward. Three coins fell. He cheered.

The Cromptons Penny Falls was first manufactured in 1964 in Ramsgate, with a refined version released by 1966. Decimalisation in 1971 retooled the slot. The euro changed nothing because it never came. The shift from one penny to two pence to whatever fractional unit will replace cash entirely has been, for this object, a series of cosmetic tweaks to a mechanism that was finished sixty years ago. Alan Meades, who wrote a social history of the British amusement arcade, calls them pivotal, the machines that, alongside the fruit machine, kept arcades solvent through the collapse of the seaside holiday and everything that came after it.

What's strange about the coin pusher is not the survival itself but the absence of any pressure to replace it. The software industry I work in cannot tolerate a system that hasn't been rewritten in three years. The financial system cannot tolerate physical currency at all if it can be helped. Yet a sweeping shelf in a Blackpool arcade, manufactured the year of Goldfinger, is still earning its keep. Nobody has built a better coin pusher because nobody needs to. The mechanism is correct. It performs the function exactly. The only thing it had to adapt to was the denomination of the coin.

The wider arcade is more layered than this. Penny pushers share floor space with light-gun shooters from the eighties, crane grabbers whose grip strength is famously calibrated to fail, and pre-decimal "old penny arcades" that have repositioned themselves as heritage attractions, charging entry to mechanical fortune-tellers and execution dioramas built between the wars. The original pioneer of all this was a Leeds mechanic named John Dennison, who started making working models in 1875 and supplied Blackpool Tower with around fifty machines that ran on its upper floors until the late sixties. Three of his daughters, Evelyn, Florence, and Alice, kept the business going. Alice did the mechanics. Most of what they built has been lost.

The point is not that the arcades are sad now. They are not. A wet Tuesday in October at Coral Island, Blackpool, is still a functioning piece of infrastructure for a child with a paper cup of two-pence pieces. The point is that almost nothing else in British public life has been allowed to persist on its own terms this long. Libraries get rebranded. Pools get demolished. Post offices close. The arcade survives because it was never institutionally important enough to be rationalised. Nobody was ever going to commission a five-year strategic review of what coin pushers are for.

Cromptons is still based in Kent, the same county the original prototype came out of. Coin pushers, in slightly varied cabinets, are still being sold. The mechanism is older than most of the people who built the rest of the seaside, and it does not appear to be going anywhere.

Sources:

Darkness, Translated

The English translation of In Praise of Shadows appeared in 1977, published by Leete's Island Books with a foreword by the architect Charles Moore, who wrote that the essay "could change our lives." By the mid-1980s it had become something very specific: the canonical text Western designers and architects reached for when they wanted theoretical cover for dimmer rooms. This is not a complaint. It is just an accurate description of what happens when an essay written in Japanese in 1933 crosses into a culture that is, at that exact moment, running out of patience with the Modernist interior.

Tanizaki's argument is simple enough in outline. Japanese architecture organises itself around shadow. The deep roof overhang of a traditional house keeps the interior dim. Lacquerware, gold embroidery, cloudy paper, their beauty depends on low light, on the way darkness lets a material breathe without the scrutiny of full exposure. Westerners, he argues, are congenitally committed to brightness: from candle to oil lamp to gaslight to electric light, their progress is a history of abolishing shadow wherever it pools. Japanese culture accommodated the dark. Western culture declared war on it.

What made this legible to an 1980s Western audience was partly timing. The translation arrived just as postmodern architecture was trying to argue its way out of the glass box. Moore himself was building Piazza d'Italia in New Orleans. The appetite for any text that questioned the clean-bright-functional dogma was considerable. Tanizaki, dead since 1965 and safely historical, offered an aesthetic that could be absorbed without political risk, and the quality he described, that particular stillness of a dim room with tatami underfoot and light filtered through shoji, landed precisely where Western designers needed it to land.

The irony that the 1980s West almost entirely missed is geographic. The same decade that architects were annotating their copies of In Praise of Shadows, Tokyo was building the city that inspired Ridley Scott's Blade Runner. Japan had become the global centre of consumer electronics: Sony, Honda, Sharp, the robotics movement, a neon-saturated culture that made the Pacific Rim feel like a preview of everywhere else's future. Tanizaki's 1933 elegy for pre-modern shadow had become, within fifty years, a description of the very country that had most completely abandoned what he mourned. Western readers, discovering the essay in translation, were perfectly positioned not to notice.

There is also a reading problem the 1980s West missed almost entirely. Tanizaki was not writing a sincere treatise. He was writing a zuihitsu, a Japanese prose form that permits fiction, irony, and digression, using a narrator who ventriloquises the tropes of cultural superiority that circulated freely in 1930s nationalist Japan. The essay's reader in 1933 Japan understood this. The Western reader in 1985, lacking that context, took the East/West binary at face value, lifted it into their own anti-modernist argument, and used it as evidence.

This is what the essay became in Western hands: permission. Permission to use wood instead of aluminium. Permission to keep a room dim. Permission to frame slowness and imperfection as values rather than deficits. All of that was probably overdue. I am not sure the fact that Tanizaki was doing something more complicated invalidates it entirely, you can extract something real from a text even when you misread it, and what the West extracted was genuinely useful. But the book it drew from is odder and funnier than the version it invented. Tanizaki admits he cannot give up electricity. He praises a toilet as a place for quiet contemplation and then spends two paragraphs on miso soup. The serene, quietly authoritative In Praise of Shadows that entered Western architectural thinking in the eighties is cleaner than the original, which is messier and considerably more entertaining.

Sources:

Amodei at the West Wing

On Friday, Dario Amodei walked into the West Wing to meet Susie Wiles, Scott Bessent, and National Cyber Director Sean Cairncross. Asked about it afterwards, Donald Trump told reporters he had "no idea" the meeting had taken place. That is, so far, the only direct presidential comment on the record.

The White House readout was more conventional: "introductory, productive, and constructive," covering "opportunities for collaboration" and "shared approaches and protocols to address the challenges associated with scaling this technology." Both Politico and CNBC confirmed the attendee list before the meeting happened. The story is not that it happened. The story is what the administration is now doing in public about a company it has spent two months calling a national security risk.

Anyone following this has the sequence memorised by now. February: Pete Hegseth tries to force Anthropic to drop its bans on autonomous weapons and mass surveillance. Amodei refuses. The Pentagon declares the company a "supply chain risk", federal agencies get a phase-out order, OpenAI signs the contract instead. March: Anthropic sues the DOD. April 7: Claude Mythos ships under Project Glasswing, restricted to about fifty partners because its zero-day-finding capability is judged too dangerous for general release. April 10: Bessent and Jerome Powell summon five bank CEOs to discuss Mythos. April 16: Bloomberg reports OMB is setting up a framework to let federal agencies use the tool. April 17: the West Wing meeting.

What's new is that the meeting is openly about routing access. The Next Web's read, and CNBC confirms the rough shape, is that any deal exits through civilian agencies and explicitly does not include the Pentagon. That is a peculiar compromise. The blacklist isn't being lifted. The supply-chain designation isn't being rescinded. Hegseth is still, nominally, correct about Anthropic being too dangerous. Treasury, CISA, and the intelligence community are simply going around him, with the Chief of Staff in the room to make it official.

One detail from CNBC is worth keeping. Wiles was previously at Ballard Partners, the lobbying firm Anthropic hired immediately after the February designation. I don't think this is a scandal , hiring lobbyists with access is what companies do when the government cuts them off, but it does change the texture of "introductory, productive, and constructive." Those are words chosen by someone who already knew the room.

The harder question is what Anthropic is actually getting. The civilian-only carve-out leaves the Pentagon's objections technically intact. The lawsuit is still live. The phase-out is still policy. What's on offer is something like conditional rehabilitation: you stay blacklisted where it began, you get sold back to everyone else, and the administration doesn't have to admit the original call was wrong.

Amodei has, in some sense, won. The product the Pentagon banned is the same product Treasury is trying to procure. The company the president said he would never do business with again had its CEO in the West Wing six weeks later. The safety features that got it blacklisted are the same safety features Treasury is asking to audit in the hope of finding flaws it can use.

And asked about any of this, the president, on the record, says he had no idea it was happening. A ten-minute briefing on a frontier AI model should probably have reached him by Friday afternoon. It is not clear that it has. It is not clear that it will.

Sources:

Maine Gets There First

On Tuesday the Maine legislature passed LD 307, a temporary ban on new data centers drawing twenty megawatts or more. The House voted 79 to 62, the Senate 21 to 13, both on party lines with a handful of exceptions. Governor Janet Mills has ten days to sign, veto, or let it become law by inaction. If she signs, Maine becomes the first US state to pause construction on what the industry calls hyperscale facilities, and what everyone else has started calling AI data centers.

Twenty megawatts sounds like a lot until you look at what a contemporary data center actually pulls. The Regional Plan Association pegs the current average at around forty. The early ones used two. So the Maine threshold isn't a ceiling on something exotic; it's a line drawn roughly at the median. The bill walks the state back to a posture from fifteen years ago.

The politics are messier than the vote count suggests. Mills is running in a contested Democratic primary for a US Senate seat and has signalled hesitation about signing without a carveout for a proposed 82MW facility in Jay. Tony McDonald, the developer, has argued his project is the wrong target, not a hyperscaler chasing AI training loads, just a large tenant that got caught in a dragnet. He may be right. He may also lose anyway, because the dragnet is the point.

What makes Maine interesting isn't the substance of the bill, which is mild. Eleven other states tried something similar in the last year and stalled or failed outright, Georgia, Maryland, Michigan, New Hampshire, New York, Oklahoma, South Carolina, South Dakota, Vermont, Virginia, Wisconsin, per Gizmodo's count. Maine's version passed partly because Maine barely has any data centers yet, the first large one is still under construction at the old Loring Air Force Base , so the industry hadn't staked out enough ground to defend.

And it passed while the federal direction runs the other way. Trump's December executive order explicitly warned that excessive state regulation thwarts the AI race, with threats to withhold funding from states that restrict growth. That's the same playbook I covered in the broadband money fight , use federal dollars as leverage against state AI laws. Maine will be a test of whether that threat actually bites a small New England state with a pipeline of maybe two data centers.

The Sachs argument, delivered in plain committee language, was that the tradeoffs haven't been shown to benefit ratepayers, water use, or community economic activity. That's a boring sentence until you remember that Maine already ranks fourth in US electricity prices. A hyperscaler sized at forty megawatts is a quiet permanent tax on every household on the grid, which Turley at OpenAI basically conceded when he admitted unlimited AI plans are like unlimited electricity plans.

The moratorium ends November 2027. A new Data Center Coordination Council gets the interval to write policy recommendations. By the time it reports, the federal posture may have shifted again, or it may have hardened. Either way, Maine's bet is that the study period is worth more than the projects it blocks. Given the two projects on the table, that math isn't hard.

Sources:

Ginny Danbury, 1989

Peter Weir shot the Midsummer Night's Dream sequence of Dead Poets Society on location in Delaware over a month in late 1988. Lara Flynn Boyle was eighteen, cast as Ginny Danbury, Chet's sister, with a second role as Hermia in the staging that anchors the film's middle act. She filmed her Ginny scenes. She filmed Hermia. She went home and waited for the release.

Then she watched the film with her mother, Sally. "Here comes my scene, here comes my scene. No scene, no scene." That's what she told People in 2024. Ginny had been cut. No one called. No one wrote. She found out because the movie was playing on a television in front of her, and she was still sitting there when the credits ran.

The Danbury subplot was a full piece of scaffolding. Ginny was a friend of Chris Noel's, a third young woman in a film that otherwise pushes its women to the edge of the frame. Cut her and you cut the character who was going to give Chris someone besides Knox to talk to, the one who was meant to play Hermia, the one who might have given the film a female friendship. The Midsummer Night's Dream scenes survive because you can't drop a staged Shakespeare sequence without breaking the film's rhythm, so Boyle walks through them wordlessly, a credited extra in what was supposed to be her role.

People didn't ask the counterfactual. If Ginny had stayed at Chris's screen weight, Boyle would have been the breakout young actress of summer 1989. Peter Weir films have a habit of making careers. Ethan Hawke's did. Robert Sean Leonard's too, for the short run he had left. Josh Charles went quiet for a decade and came back to television. Instead, Boyle got the cleanest dispatch Hollywood issues: a credit on a film that contains almost none of her.

She was already familiar with the shape of the edit. Ferris Bueller's Day Off had cut her three years earlier. Dead Poets made her two-for-two.

A year after watching her mother's screen skip over her, she was in Washington state playing Donna Hayward. "Twin Peaks gave me everything I have as an actor," she said. It's hard to picture the timeline where Ginny survives the edit and Lynch still casts her. There aren't many universes where the 1989 breakout of a Robin Williams vehicle signs on to a strange little ABC pilot about a murdered homecoming queen.

Sources:

Cut Along the Dotted Line

The coupon was usually in the bottom corner of the right-hand page, bordered with a dashed line and the instruction to cut here. You cut. You filled in your name and address in capital letters. You walked to the post office with a parent or on your own if you were old enough, stood in the queue, and asked for a postal order for one pound ninety-nine. The woman behind the grille filled in the amount, stamped it, and handed it across with a receipt. You folded the order into an envelope with the coupon, stuck on a second-class stamp, and dropped it into the pillar box on the way home. Then you waited.

The waiting is the part that has become unrecoverable. Most small ads specified "please allow 28 days for delivery," and that figure was realistic rather than padded. The ad you'd torn out lived in Exchange & Mart, or the back pages of Roy of the Rovers, or somewhere inside Smash Hits. The company processing your order might be one man operating out of a garage. Your postal order had to clear. The catalogue had to be printed. Stock had to be located. The padded envelope came back eventually, franked by a post office hundreds of miles away, the address scrawled in a handwriting you'd never seen.

In the month between sending and arriving, the object lived entirely in your head. X-ray specs that saw through skin. Sea monkeys that performed synchronised dances. A magic set whose tricks the advert had strongly implied would fool everyone. The mental image grew more specific and more extraordinary the longer the padded envelope took. You could not check on its progress. There was no tracking number, no notification of dispatch, no photograph of the warehouse worker who'd packed it. The order entered a system and disappeared from view, and your imagination filled the silence.

When the padded envelope arrived, the object inside could not win. The X-ray specs were pieces of plastic with cardboard lenses that made everything look striped and red. The sea monkeys were brine shrimp that hatched to roughly the size of a comma. The magic set came with trick cards you could see through in good light. You held the thing in your hand and felt the distance between the copy in the advert and the object the padding had protected. Then you played with it for an afternoon and mostly forgot.

Nostalgia is not the right register for any of this. The objects were nearly always a let-down. What has disappeared is not the stuff but the structure of expectation the stuff was suspended inside. A whole month of specific, named waiting, knowing exactly what you'd ordered and unable to retrieve or cancel or check. The desire had time to become baroque before meeting reality.

The small ads themselves have mostly vanished from regional press. Their mail-order equivalents migrated to online storefronts that list, illustrate, price, and review everything in one page without requiring you to tear anything out of anywhere. The padded envelope has been replaced by the brown Amazon box, which arrives on a schedule you can track hour by hour. That the product inside is often the same plastic tat imported from the same factories is beside the point. Next-day delivery cannot sustain the same quality of imagining. The real object arrives before the imagined one has started to grow.

Plastic tricks and broken toys still turn up in charity shops and at car boot sales, slipped from their time but intact. The X-ray specs outlast the magazine that advertised them. Someone ordered them in 1983 and kept them in a drawer. The padded envelope is long gone, but the object still carries the shape of something that was waited for. An artefact that came through the post moves on differently from one that came over a counter. It always contained an interval.

A postal order can still be bought at any British post office. I don't know anyone under forty who has ever held one.

Sources:

Palette Icon, Canva Engine

On Friday, Anthropic launched Claude Design, the AI design tool I wrote about on Wednesday when it was still a single-source scoop in The Information. The product is now live as a research preview for Pro, Max, Team, and Enterprise subscribers, reachable via a palette icon on Claude.ai's left-hand navigation. It is powered by Opus 4.7. Figma dropped about 7% on the news. The market made the same mistake twice. What actually shipped is neither of the two shapes I expected.

The question I raised on Wednesday was whether Claude Design would be hosted-only or API-accessible. A hosted product is a controlled experiment. An API-accessible one would reset the wrapper economy , the cohort of Lovable, Bolt, v0, Cursor builders who pipe prompts into frontier models and ship a UI on top of the response.

Anthropic shipped neither version cleanly. The product is hosted only, no API. But it isn't running on design infrastructure Anthropic owns. It runs on Canva's Design Engine. Output is exportable as PDFs, URLs, or PowerPoints, or handed off directly to Canva for drag-and-drop editing. Canva's own Create event was in Los Angeles the same day, where they announced Canva AI 2.0 and called it the biggest product launch in company history. Both launches were timed together, which is not how opportunistic partnerships work.

This is a third shape I didn't account for: a hosted-only product that gets reach through somebody else's installed base. Canva has distribution inside enterprise marketing teams. Anthropic does not. The partnership gives Anthropic that distribution without building it, and gives Canva a direct pipe into the frontier-model side Adobe has been slow to acquire. It's joint, not cannibalistic.

Which partially inverts the thesis. I argued Figma and Adobe were the wrong targets. That's still true; they trade in coordination and indemnified enterprise contracts Claude Design doesn't touch. I argued Lovable, Bolt, and v0 should be the ones panicking. The Register named Lovable explicitly as the shot-across-the-bow target. That half lands as predicted.

What I didn't see coming: Canva is the structural winner. The consumer-grade design generators (Microsoft Designer, Adobe Express's free tier) lose more than Figma ever was going to, because they don't have Opus sitting underneath their prompt box and they don't get the dual announcement billing. Opus 4.7 itself shipped a day earlier, framed in Anthropic's own launch materials as "less broadly capable" than the restricted Mythos. That framing makes sense now. Opus 4.7 is the commercial frontier with Claude Design parked inside it. Mythos is the internal model kept in reserve. The two launches rhyme in a single week because they are solving different problems in the same portfolio.

The interesting question is when the API arrives. A hosted-only Claude Design is a controlled experiment. A Canva-backed Claude Design is a distribution play. A Claude Design with an API, and maybe a Canva-rendered API, would be the real wrapper-layer displacement. None of the coverage I've read this morning says anything about API access. Anthropic's research-preview framing keeps all of that future-tense.

The signal I was waiting for on Wednesday has shifted. It's no longer "what does Anthropic do about the wrapper economy." It's "what does Anthropic do with Canva six months from now."

Sources:

Counting R's in Strawberry

Ask a frontier model to count the letter 'r' in "strawberry." Often you'll get two. Sometimes three. Rarely with consistent accuracy across a hundred trials. This isn't a bug somebody hasn't fixed. It's a direct consequence of how text enters the model in the first place, and it explains a dozen other strange behaviours that cluster around counting, spelling, and arithmetic.

Language models don't read text. They read tokens, which are integers that index a lookup table. The algorithm most of them use to build that table is called Byte-Pair Encoding, or BPE. It was originally designed as a data compression scheme. Philip Gage wrote it in 1994 for compressing files. Sennrich, Haddow, and Birch adapted it for machine translation in 2016, and it's been the default across most major model families since.

The idea is simple. Start with individual characters. Find the most frequent adjacent pair in your training corpus. Merge them into a new token. Find the next most frequent pair. Merge. Repeat for tens of thousands of iterations. The result is a vocabulary full of common words and common subword pieces. "The" is one token. "ing" is one token. "Strawberry" might split into ["straw", "berry"]. Rare words fragment into more pieces.

Once the model is trained, it sees "strawberry" as two integers, not ten characters. No mechanism inside the transformer can reach inside a token to ask how many r's it contains. The letters are sealed inside the token the way pages are sealed inside a book you can only see the cover of. The model has, statistically, learned that strawberry contains three r's. It just hasn't learned it from the token sequence. It's learned it from surrounding text that happened to mention the fact. That knowledge is fragile, and it decays on uncommon words.

The arithmetic failures come from the same place. Numbers don't tokenize uniformly. "480" might be one token. "481" might be two. A four-digit number can split one way and the same digits rearranged can split another. Researchers using arithmetic as a diagnostic have found that when an answer has more digits than either input, accuracy on certain tasks collapses to under 10%. The model isn't bad at maths. It's being handed digit sequences in a shape it wasn't trained to work with.

The fix, in principle, is byte-level tokenization. Every byte becomes a token. No merging, no hidden letters. The tradeoff is sequence length. The same passage takes more tokens, sometimes many more. That means more compute, longer context windows, slower inference. GPT-2 went byte-level and paid the cost. Recent models use hybrid approaches: BPE for efficiency, special handling for digits, sometimes per-character processing inside the chain-of-thought. None of it is free.

What strikes me is how much of the model's apparent cognitive style is downstream of this one preprocessing choice. The blind spot for letters isn't a reasoning failure. It's the data format not including letters, most of the time. Change the tokenizer and you change what the model can notice.

The cost-per-task shift playing out across the industry is partly a reasoning-modes story, but it's also a tokenization story. The reasoning variants often break digits into individual character tokens during arithmetic steps, which inflates token counts for the same sum. The reason your cheap tier struggles with long division is the same reason your expensive tier's bill keeps climbing when you ask it to do the job properly. Both are paying, differently, for the fact that the default encoding hides the thing you want the model to count.

Sources:

Eight Years, No Walkout

Google is negotiating with the Department of Defense to let the Pentagon run Gemini on classified networks, The Information reported on Thursday. Reuters, Engadget, and a handful of others picked it up the same day. The proposed contract reportedly carves out two exclusions: no mass domestic surveillance, no autonomous lethal weapons. It is, in shape, the OpenAI deal. That framing is the story.

What Google already runs inside DoD is larger than I realised. Since December, the GenAI.mil portal has given Gemini to around 1.2 million Defense Department staff across a user base of more than three million. Roughly forty million prompts and four million documents have gone through it. Eight pre-built agents handle work the Pentagon apparently considers administrative: meeting notes, budgets, sanity-checks against the national defense strategy. There's a feature called Agent Designer that lets personnel build their own agents in plain English. None of that is classified. It is also not nothing.

The new deal is the next step. Same infrastructure, cleared for secret and top-secret environments. The Under Secretary of Defense for Research and Engineering was quoted saying expansion talks are "underway."

Eight years ago this would not have happened. In 2018 Google pulled out of Project Maven, the drone-footage computer-vision contract, after employee protests. The company declined to renew it.

The two new exclusions, mass domestic surveillance and autonomous lethal weapons, are the same two items Anthropic refused to drop in February, when the Pentagon blacklisted them for keeping those commitments. Hegseth gave Amodei a Friday deadline; Amodei refused; the company was designated a supply chain risk. OpenAI accepted the terms Anthropic wouldn't and kept its contract. Google, eight years after walking away, is now pitching itself on roughly that middle ground.

It is a narrower position than 2018 Google held. It is a wider position than February Anthropic held. In the current window, it is where the business lives.

The part I keep circling is the silence. In 2018 the Maven protest was a company-wide story with a visible fracture line and a public exit. This week's news is a Reuters summary citing an Information scoop, picked up through the wire services, noted by industry press. There's no internal petition making the rounds. No engineers are speaking anonymously to the Times. The deal and its terms are being negotiated in the normal way, which is to say without anyone getting in the way of it.

That might be because the workforce has changed. It might be because what looked plainly wrong in 2018, helping the military see, has been reclassified as ordinary productivity software that happens to have some optional national security use cases attached. The agents summarising budgets look very much like the ones Anthropic was meant to be, before the February argument.

I don't think this is the same company that walked away from Maven. The contract terms say something Google once said it wouldn't sign. The absence of a visible fight says the company isn't planning to argue about it.

Sources:

Rosalind Without the Promises

OpenAI released GPT-Rosalind yesterday, the first entry in a new "Life Sciences" model series, gated to a trusted-access programme for a handful of enterprise partners. Amgen, Moderna, Thermo Fisher, the Allen Institute, and Los Alamos National Laboratory are on the preview list. Named after the crystallographer who made DNA legible, the model ships in ChatGPT Enterprise, Codex, and the API, behind enterprise-grade security controls and the standard "no training on your data" clause.

The benchmark numbers are good. On BixBench, GPT-Rosalind scores 0.751 Pass@1. On LABBench2, it wins six of eleven subtasks. Against human experts on two representative tasks, it sits in the 95th and 84th percentiles. Reasonable results for a domain model built on top of the frontier reasoning stack.

But the interesting thing isn't the benchmarks. It's the language around them.

Read the announcement carefully. OpenAI doesn't say GPT-Rosalind will design drugs. The model is described as a tool to accelerate the early stages of discovery, evidence synthesis, hypothesis generation, experimental planning. That's the research-assistant frame, not the autonomous-designer frame. The Codex Life Sciences plugin talks to more than fifty scientific databases. The model reads papers, cross-checks datasets, drafts experiments. That is useful. It is not a cure.

Contrast that with the last five years of pure-play AI drug discovery. Exscientia, Recursion, BenevolentAI, the three companies that raised the most money on the premise that AI could find drugs faster, have all had their first clinical readouts. All three were negative. Recursion absorbed Exscientia. BenevolentAI has been in retrenchment. The sector is sitting on roughly a billion and a half in market cap and has approved zero drugs. UCL's Peter Coveney, cited in a recent Nature piece, has made the structural case: discovery isn't the bottleneck. Validation is. You can generate ten thousand candidate molecules. Testing them is the part that takes a decade.

GPT-Rosalind isn't promising to solve validation. It's promising to make the scientists who do validation a little faster at reading papers. That's a smaller claim. It might also be a correct one.

There's something honest about the framing. Opus 4.7 shipped yesterday with Mythos held back, carefully gated, the same instinct in a different domain. OpenAI's move here rhymes: trusted-access only, enterprise-only, a short partner list, no individual researchers, dual-use safety language that reads as if somebody who has read the risk literature wrote the announcement.

Whether this is restraint or marketing discipline is open. It's possible GPT-Rosalind is simply a harder model to sell on hype because pharma buyers have been burned too many times and know better. It's possible the trusted-access structure is there to keep the model from generating plausible-but-wrong bioinformatics claims at industrial scale.

Either way, it's the first new model family in a while whose launch language reads like it was written by people who watched the AI-for-biology narrative play out in public. Rosalind Franklin did careful work and died before seeing it credited. Putting her name on a frontier model is either a very nice gesture or a warning about what happens when you overstate.

Sources: