This is going to be the first in a series of articles on
liquid cooling in the data centre environment, with a view to sparking some
debate.
The use of liquid cooled technologies in the data centre has
been akin to debates about nuclear fusion, insofar that its always 5-10 years
from adoption.
We've been tracking the use of liquid cooled technologies
for some time now and we are party to some developments that have been taking
place recently, and it is my belief that we are going to see some serious
disruption taking place in the sector sooner rather than later.
We'll be looking at a unique way of looking at how to implement liquid cooled solutions in the next few posts, but before that we need to understand the direction of
travel outside of the DC Space.
Customers are increasingly using cloud services
and it appears that organisations that are buying physical equipment are only
doing so because of either "thats the way we've always done IT" or
that they are using specialist applications that cannot be provide via cloud
services or they are wedded to some out of date procurement process. This is
supported by the fact that conventional off the shelf server sales are in
decline, whilst specialist cloud enabled servers on on the up but being
purchased by Cloud Vendors and the Hyperscalers (albeit that they are designing
and building their own servers which by and large are not suitable for
conventional on premise deployements) There is also the Open Compute Project
that is mostly being used for pilots via development teams.
So, with cloud servers, the customer, in fact any customer
has no say in the type of physical hardware deployed to provide that SaaS, PaaS
or IaaS solution and that is the right approach. IT is increasingly becoming a Utility,
just like Electricity, Water and Gas. As a customer of these utilities I have
no desire to know how my electricity reaches my house, all I want is that when
I flick the light switch or turn on the TV, that the light comes on and that I
can watch the "Game of Thrones" I certainly do not care which power
station generated the electricity, how many pylons, how many transformers, how
many different cables etc that the "juice" passed through to get to
my plug/switch.
For digital services, I access a "service" on my
smartphone, tablet, desktop etc and via a broadband or wifi connection, access the "internet" and then route through various buildings
containing network transmission equipment to the server (which can be located
anywhere on the planet), the physical server(s) that the "service"
resides upon. Everything, apart from my own access device is not my asset, I
merely "pay" for the use of it. And I dont pay directly, I pay my ISP or Internet Service Provider a monthly fee for access and the supplier directly for the service that I am accessing and somehow they pay everybody else.
So, in essence, digital services are (to me) an app I select on my digital device, information is then sent and received over digital infrastructure, either fixed such as a broadband connection or via a wifi network. I have no knowledge or, nor do I care how the information is sent or received, merely that it is.
What does this have to do with the use of air or liquid in a data centre?
Well, its actually quite important, but before we cover that, we'll be covering the basics of data centres and we'll be doing that in the next post.