What’s the distinction between a silly thought and a superb one? Typically, it simply comes right down to sources. Virtually limitless funds, like limitless thrust, can get even a mad thought off the bottom.
And so it may be for the idea of placing AI data centers in orbit. In a uncommon second of unalloyed settlement, a few of the richest and strongest males in expertise are staunchly backing the thought. The group contains Elon Musk, Jeff Bezos, Jensen Huang, Sam Altman, and Google CEO Sundar Pichai. In all probability, a whole bunch of individuals at the moment are engaged on the idea of area knowledge facilities on the corporations immediately or not directly managed by these males—SpaceX, Starlink, Tesla, Amazon, Blue Origin, Nvidia, OpenAI, and Google, amongst others.
Probably prices to design, construct, and launch a 1-GW orbital datacenter, primarily based on a community of some 4,300 satellites and together with working prices over a five-year interval, would exceed US $50 billion. That’s about thrice the price of a 1-GW knowledge middle on Earth, together with 5 years of operation.John MacNeill
So how a lot would it not price to begin coaching large language models in area? Most likely the most effective accounting is one created by aerospace engineer Andrew McCalip. McCalip’s exhaustive, detailed evaluation contains interactive sliders that allow you to examine prices for space-based and terrestrial knowledge facilities within the vary of 1 to 100 gigawatts. One-gigawatt data centers are being built now on terra firma, and Meta has introduced plans for a 5-GW facility, with anticipated completion a while after 2030.
In an interview, McCalip says his preliminary tough calculations a number of years in the past urged that knowledge facilities in area would price within the vary of seven to 10 occasions extra, per gigawatt of capability, than their terrestrial counterparts. “It simply wasn’t sensible,” he says. “Not even shut.” However when Elon Musk started publicly backing the thought, McCalip revisited the numbers utilizing publicly out there details about Starlink’s and Tesla’s applied sciences and capabilities.
That modified the image considerably. The figures in his on-line evaluation assume an orbital community of data-center satellites that borrows closely from Musk’s tech treasure chest—“primarily…you simply begin placing some radiation-resistant ASIC chips on the Starlink fleet and also you begin rising edge capability organically on the Starlink fleet,” McCalip says. The community would depend on the form of watt-efficient GPU architecture utilized in Teslas for self-driving, he provides. “You begin dropping these onto the backs of Starlinks. You possibly can slowly develop this out, and this could be roughly the efficiency that you’d get.”
Backside line, with some stable however not essentially heroic engineering, the price of an orbital knowledge middle may very well be as little as thrice that of the comparable terrestrial one. That differential, whereas nonetheless excessive, at the least nudges the idea out of the immediately dismissible class. “I’ve my explicit views, however I would like the information to talk for itself,” McCalip says.
For this illustration, we picked a configuration with an combination 1 GW of capability. The community would encompass some 4,300 satellites, every of which might be outfitted with a 1,024-square-meter photo voltaic array that generates 250 kilowatts. The info middle on that satellite tv for pc, powered by the array, may need at the least 175 GPUs; McCalip notes {that a} well-liked GPU rack, Nvidia’s NVL72, has 72 GPUs and requires 120 to 140 kW.
The whole price of the satellite tv for pc community could be round US $51 billion, together with launch and 5 years of operational bills; a comparable terrestrial system would price about $16 billion over the identical interval.
Silly? Not silly? You resolve.
From Your Website Articles
Associated Articles Across the Internet

