Is there an ideal place for every workload?

The idea that everyone has a perfect job, or niche, based on their personality and skill set could either be seen as a positive or potentially constraining. My latest blog over at Verne Global applies that idea to computing and looks at the idea of Best Execution Venue for workloads and whether there’s really an ideal venue for every workload to thrive.

Supercomputers in the cloud erode another case for owning data centers

My latest column over at Data Center Knowledge looks at supercomputing in the cloud.

HPC is obviously a nebulous term that covers a whole range of systems from mid-level servers right up to supercompute. The disruption from cloud is probably being felt most at the lower end but it’s still notable that a supercomputer supplier like Cray is now plugged into Azure.

Screen Shot 2017-10-30 at 09.33.10The large scientific supercomputing institutions are unlikely to farm our workloads to cloud but there’s likely to be appetite from industry as demand for deep learning in fields such as automotive and aerospace ramp up.

Why the Data Center Liquid Cooling Space is Heating Up

Screen Shot 2017-10-20 at 13.59.42Latest column for Data Center Knowledge is on direct liquid cooling (DLC). I’ve been researching and writing about DLC for several years now but it finally feels like there is demand-side momentum building to match the fervour from suppliers. The pros and cons of the various flavours of DLC still stand but with support from all of the main server OEMs, adoption by some hyperscales, and even interest from colos, liquid cooling may be finally reaching a tipping point beyond traditional HPC.

Data Centers Will Invest in Renewables Despite the EPA, Not Because of It

Screen Shot 2017-10-12 at 08.13.27

This week’s column for Data Center Knowledge looks at the EPA’s recent decision to repeal the troubled Climate Power Plan. A number of large hyper-scale operators have come out in support of the plan which they claim would help them, and wider industry, with future renewable investments.

It has been pointed out that these hyper-scales invest in renewables because of tax incentives to do so rather than out of a sense of altruism. That is accurate up to a point. The hyperscale operators admit themselves that ““Renewable energy is less subject to price volatility than non-renewable energy; provides greater long-term cost certainty to its purchasers; and, in many parts of the United States, is available at prices comparable to or better than the current prices for other electricity options”. But as Greenpeace – which has been highly critical of some datacenter operators use of renewables –  and others have pointed out some large operators have helped to make renewable energy tariffs more widely available and increase use of renewables by other industries.

There is a lot of complexity around the sourcing and distribution of renewables but the decision by the EPA to roll-back the Climate Power Plan still looks like a retrograde step.

AI and Edge Could Usher in a New Era of Liquid-Cooled Data Centers

I recently spoke with Tate Cantrell the CTO over at Verne Global about his views on direct liquid cooling in datacenters. The article is posted on Verne’s site. Tate, like many other datacenter technologists, tends to take a pragmatic view of liquid cooling but says the colocation provider is happy to accommodate DLC if it’s what the customer wants.

There are good reasons to think that more customers will want to use DLC technology as high-density workloads associate with machine learning and deep learning proliferate. That’s good news for the 20-plus suppliers of specific DLC technology as well as all the main server OEMs which have licensed that technology or developed their own in-house. Some edge deployments – especially ruggedised, non-whitespace micro-modular designs – could also be good candidates for DLC and several suppliers are also targeting this use case. Screen Shot 2017-10-09 at 10.44.35

Datacenter resiliency: It’s the network, stupid!

Screen Shot 2017-10-06 at 11.19.19

My latest column over at Data Center Knowledge picks through some of the latest thinking on datacenter resiliency from experts such as Uptime Institute as well as the actual resiliency strategies deployed by some cloud service providers.

The upshot is that some cloud service providers are investing heavily in new approaches to IT and datacenter resiliency based on the use of distributed, virtualized applications, instances, or containers using middleware, orchestration, and distributed databases across multiple datacenters.

“The evolution of compute from single-site data centers with proven in-house engineering to a multi-site hybrid computing model is enabling distributed models of resiliency that can achieve high availability and rapid recovery alongside improved performance,” Uptime Institute stated in a recent webcast. “The recent move by many to the public cloud is further accelerating this cloud-based resiliency approach. The benefits are potentially vast.”

Now, an easy conclusion to jump to would be that improved resiliency at the IT level should mean that the importance of individual datacenter is lessened.  As such it should be possible to design and build new facilities with less redundant M&E equipment: fewer generators, UPS etc. However, while that might be possible in some instances – perhaps in edge datacenters – the reality is that most operators continue to build to Uptime Institute Tier III or equivalent (sorry Uptime I know there is no real equivalent).

As the outage this week in Microsoft’s Azure’s service shows even cloud service providers with advanced resiliency in place – Microsoft recently introduced Availability Zones – aren’t immune to downtime. True, Microsoft said that “customers with redundant virtual machines deployed across multiple isolated hardware clusters would not have been affected by the outage,” but it seems some customers were affected.

The take-away is that for now advanced approaches to redundancy – based on software and networks – should probably viewed as an additional “belt” to support the existing “braces” of conventional single-site redundant M&E and IT.

For more on the issue check out this very good webcast overview of Uptime’s current thinking on advanced resiliency.