Was looking through my office window at the data closet and (due to angle, objects, field of view) could only see one server light cluster out of the 6 racks full. And thought it would be nice to scale everything down to 2U. Then day-dreamed about a future where a warehouse data center was reduced to a single hypercube sitting alone in the vast darkness.
They look silly now. Many data centers are not scaling up power per rack. With GPUs, there are often two chassis per rack.
Have that problem ourselves, they didn’t provision power or cooling for this kind of density, and how do you pipe in multiple megawatts to a warehouse in the middle of nowhere?
Only if storage density out paces storage demand. Eventually, physics will hit a limit
Physics is already hitting limits. We’re already seeing CPUs be limited by things like atom size, and the speed of light across the width of the chip. Those hard physics limitations are a large part of why quantum computing is being so heavily researched.
You think that if we can scale 6 racks down into one cube that someone wouldn’t just buy 6 racks of cubes?
They’ll always hunger for more.
I think what will happen is that we’ll just start seeing sub-U servers. First will be 0.5U servers, then 0.25U, and eventually 0.1U. By that point, you’ll be racking racks of servers, with 10 0.1U servers slotted into a frame that you mount in an open 1U slot.
Silliness aside, we’re kind of already doing that in some uses, only vertically. Multiple GPUs mounted vertically in an xU harness.
You’ve reinvented blade servers
The future is 12 years ago: HP Moonshot 1500
“The HP Moonshot 1500 System chassis is a proprietary 4.3U chassis that is pretty heavy: 180 lbs or 81.6 Kg. The chassis hosts 45 hot-pluggable Atom S1260 based server nodes”
Yeah, that’s the stuff.
That did not catch on. I had access to one and the use case and deployment docs were foggy at best
It made some sense before virtualization for job separation.
Then docker/k8s came along and nuked everything from orbit.
The other use case was for hosting companies. They could sell “5 servers” to one customer and “10 servers” to another and have full CPU/memory isolation. I think that use case still exists and we see it used all over the place in public cloud hyperscalers.
Meltdown and Spectre vulnerabilities are a good argument for discrete servers like this. We’ll see if a new generation of CPUs will make this more worth it.
128-192 cores on a single epyc makes almost nothing worth it, the scaling is incredible.
Also, I happen to know they’re working on even more hardware isolation mechanisms, similar to sriov but more enforced.