Google's Urs Hölzle explains why beefier cores are better than whimpy cores

The Register covers a new paper by Google's Urs Hölzle.

Google ops czar condemns multi-core extremists

Sea of 'wimpy' cores will sink you

By Cade Metz in San FranciscoGet more from this author

Posted in Servers, 17th September 2010 07:04 GMT

Free whitepaper – The Reg Guide to Solutions for the Virtual Era

Google is the modern data poster-child for parallel computing. It's famous for splintering enormous calculations into tiny pieces that can then be processed across an epic network of machines. But when it comes to spreading workloads across multi-core processors, the company has called for a certain amount of restraint.

With a paper (PDF) soon to be published in IEEE Micro, the IEEE magazine of chip and silicon design, Google Senior Vice President of Operations Urs Hölzle – one of the brains overseeing the web giant's famous back-end – warns against the use of multi-core processors that take parallelization too far. Chips that spread workloads across more energy-efficient but slower cores, he says, may not be preferable to chips with faster but power-hungry cores.

The paper is here and only 2 pages long.  And, when thinking what motivated Urs to write this paper i think it was his frustration that too many people are focusing on the number of cores to solve a problem and not taking into consideration what happens to the overall system when you try to solve problems with a bunch of whimpy cores vs. brawny cores.

We classify multicore systems as brawny-core systems, whose single-core performance is fairly high, or wimpy-core systems, whose single-core performance is low. The latter are more power efficient. Typically, CPU power decreases by approximately O(k2) when CPU frequency decreases by k, and decreasing DRAM access speeds with core speeds can save additional power.

Urs as usual uses excellent presentation skills to make his point in three areas.

First, the more threads handling a parallelized request, the larger the overall response time. Often all parallel tasks must finish before a request is completed, and thus the overall response time becomes the maximum response time of any subtask, and more subtasks will push further into the long tail of subtask response times. With 10 subtasks, a one-in-a-thousand chance of suboptimal process scheduling will affect 1 percent of requests (recall that the request time is the maximum of all subrequests), but with 1,000 subtasks it will affect virtually all requests.

In addition, a larger number of smaller systems can increase the overall cluster cost if fixed non-CPU costs can’t be scaled down accordingly. The cost of basic infrastructure (enclosures, cables, disks, power supplies, network ports, cables, and so on) must be shared across multiple wimpy-core servers, or these costs might offset any savings. More problematically, DRAM costs might increase if processes have a significant DRAM footprint that’s unrelated to throughput. For example, the kernel and system processes consume more aggregate memory, and applications can use memory-resident data structures (say, a dictionary mapping words to their synonyms) that might need to be loaded into memory on multiple wimpy-core machines instead of a single brawny-core machine.

Third, smaller servers can also lead to lower utilization. Consider the task of allocating a set of applications across a pool of servers as a bin-packing problem—each of the servers is a bin, and we try to fit as many applications as possible into each bin. Clearly that task is harder when the bins are small, because many applications might not completely fill a server and yet use too much of its CPU or RAM to allow a second application to coexist on the same server. Thus, larger bins (combined with resource containers or virtual machines to achieve performance isolation between individual applications) might offer a lower total cost to run a given workload.

How many data center operation VPs can write this paper?  One.  :-)

Keep the number of cores in mind for a green data center, smaller energy efficient processors may not be the most efficient overall.

Read more

Greenpeace launches Unfriend Coal Video at Facebook - the so coal network

There are some out there surprised by Greenpeace's action at Facebook, but I could see this coming 2 years ago, and even blogged incorrectly that Google, Apple, Microsoft were the potential targets in July 2009 which was logical given their brand and data center build outs.

What is the First Greenpeace Data Center Target? Apple? Google? Microsoft?

Datacenterknowledge blogs on how quickly Apple is building its $1 billion dollar data center.

Apple Moving Quickly on NC Project
July 28th, 2009 : Rich Miller

apple-nc

 

Facebook's move to Oregon and the choice of coal power set Facebook as the target.

Facebook bets on coal for new Oregon data center

By Matt Stansberry, Executive Editor
29 Jan 2010 | SearchDataCenter.com

On Jan. 21, when Jonathan Heiliger, vice president of technical operations atFacebook announced the company would build its first data center, it wasn't a surprise that the Web giant located its facility in Oregon. What is surprising is that it will not avail itself of the region's famous hydroelectric power.

This is turning into a media battle and Greenpeace launches its latest.

Greenpeace Launches Its Latest Anti-Facebook Volley

content by Greener World Media

By Matthew Wheeland at Greener World Media

Thu Sep 16, 2010 1:00am EDT

"The Social Network" it certainly isn't, but Greenpeace today unveiled the latest front in its battle to get Facebook to change its coal-powered-data-center ways.

Timed to the release in two weeks of a feature film about the at-times-seamy beginnings of Facebook, Greenpeace has put together its own satire of the company's history.

The video is here on YouTube with 34,414 views now.

Note the play on words.

image

image

and thanks to YouTube's closed caption here is some screen shots with the words.

image

image

image 

Here is a list of media coverage so far.  Maybe Facebook should have had a green data center strategy beyond being efficient? duh.

You're 'So Coal': Angling to Shame Facebook

New York Times (blog) - Leslie Kaufman - ‎1 hour ago‎

It's not intuitive to have pity for Mark Zuckerberg, the 26-year-old co-founder of Facebook. But lately Mr. Zuckerberg, thought to be the ...

30 Second MBA: Mark Zuckerberg CEO of Facebook "How do you generate innovation?"

Fast Company - ‎12 hours ago‎

Mark Zuckerberg, CEO of Facebook, answers the question: "How do you generate innovation?"

Zuckerburg would be happy to make Facebook non-profit: David Kirkpatrick

Economic Times - ‎14 hours ago‎

You ask David Kirkpatrick, why we should read his book and pat comes the reply: “If you want to understand Facebook and know who Mark Zuckerberg is. ...

Greenpeace Launches Its Latest Anti-Facebook Volley

Reuters - Matthew Wheeland - ‎21 hours ago‎

"The Social Network" it certainly isn't, but Greenpeace today unveiled the latest front in its battle to get ...

The So Coal Network: Confronting Facebook's Coal Problem (Video)

Huffington Post - ‎Sep 16, 2010‎

The story about the founders of Facebook, The Social Network, premieres next week, and Greenpeace has taken the opportunity to create our own short film, ...

Greenpeace slams Facebook's coal-powered data centre

Information Age - ‎3 hours ago‎

Greenpeace has issued a video calling on social networking giant Facebook to stop powering its Oregon data centre with electricity generated by burning coal ...

Cause for a viral video

Ithaca College The Ithacan - Rebecca Webster - ‎12 hours ago‎

It always amazes me what seemingly simple graphics and the sound of a child's voice can do for a cause. At this point, if you are an avid Facebook-er, ...

All 22 related articles »

Blogs

Mark Zuckerberg: Social Games Are Next-Generation, Xbox And Nintendo, Last-Gen

Forbes (blog) - Oliver Chiang - ‎Sep 15, 2010‎

It's no surprise that Facebook chief executive Mark Zuckerberg thinks that games, and all online applications, are becoming more social. ...

Greenpeace International Challenges Facebook's Use of Coal-Based Power

Green Jobs Ready (blog) - ‎17 hours ago‎

by Mandy on September 16, 2010 Leading non-profit environmental organization Greenpeace International is challenging Facebook to seek clean energy ...

How Green is Green Enough?

Data Center Journal (blog) - jeffrey clark - ‎20 hours ago‎

The recent (and ongoing) dispute between the social-networking giant Facebook and environmental advocacy group Greenpeace (“Greenpeace Versus Facebook”) ...

Read more

Where are Facebook's data centers? Follow the money

RichMiller at DataCenterKnowledge has a good post that answers a question of where are Facebook's data centers?

Facebook: $50 Million A Year on Data Centers

September 16th, 2010 : Rich Miller

A look at the fully-packed racks inside a Facebook data center facility.

An analysis of Facebook’s spending with data center developers indicates that the company is now paying about $50 million a year to lease data center space, compared to about $20 million when we last analyzed its leases in May 2009.

When you spend $50 million a year, you can follow the money trail.

  • Facebook is paying $18.13 million a year for 135,000 square feet of space in data center space it leases from Digital Realty Trust (DLR) in Silicon Valley and Virginia, according to data from the landlord’s June 30 quarterly report to investors.
  • The social network is also leasing data center space in Ashburn, Virginia from DuPont Fabros Technology (DFT). Although the landlord has not published the details of Facebook’s leases, data on the company’s largest tenants reveals that Facebook represents about 15 percent of DFT’s annualized base rent, which works out to about $21.8 million per year.

...

Read more

Can Google go below 1.10 PUE with Sea Water Cooling?

DataCenterKnowledge reports on Google's use of Sea Water in its Hamina, Finland data center.

Google Using Sea Water to Cool Finland Project

September 15th, 2010 : Rich Miller

Google will use cool sea water in the cooling system for its new data center in Hamina, Finland, which is under construction and scheduled to go live early next year. The initiative continues Google’s focus on data center efficiency and sustainability. Using cool water allows Google to operate without energy-hungry chillers, and also limits the facility’s impact on local water utilities.

Where there is money savings there is typically less waste.  So, will this allow Google to go below 1.10 PUE?

Here are Google's latest numbers for Q1 2010.

Q2 2010 Performance

Quarterly energy-weighted average PUE:
1.17

Trailing twelve-month energy-weighted avg. PUE: 
1.18

Individual facility minimum quarterly PUE:
1.13, Data Center J

Individual facility minimum TTM PUE*:
1.13, Data Center B

Individual facility maximum quarterly PUE:
1.22, Data Center A

Individual facility maximum TTM PUE*:
1.23, Data Center H

* Only facilities with at least twelve months of operation are eligible for Individual Facility TTM PUE reporting

For more details on Google's latest data center construction.

The company’s plans were discussed in an article in Computer Sweden (translation inEnglish), which got a tour of the construction site in Hamina. There are no servers in sight yet, but the story reports that Google has refurbished the water pumps used at the former newsprint plant, and will use large pipes to draw cool water from the nearby Baltic Sea.

Google has a great goal for reducing its water consumption.

Google hopes to eventually use recycled water for up to 80 percent of the company’s total data center water consumption. “The idea behind this is simple: instead of wasting clean, potable water, use a dirty source of water and clean it just enough so it can be used for cooling,” Google says on its water management web page. “Cooling water still needs to be processed, but it’s much easier to treat it enough for data center use compared to cleaning it for drinking use.”

Read more

Environmental Impact of Data Center Diesel Generators - Quincy, WA

DataCenterKnowledge has a post about Washington State's Department of Ecology studying the environmental impact of Data Center Diesel Generator operation in Quincy, WA.

Quincy Generator Cluster Draws Scrutiny

September 13th, 2010 : Rich Miller

An aerial view of the Microsoft data center in Quincy, Washington

Economic development officials love clusters of huge Internet data centers. But environmental officials are less enthused about large clusters of diesel generators. The town of Quincy, Washington has both, serving as home to major data centers for Microsoft, Yahoo and Intuit (with another project from Sabey Corp. on the way). 

As much as the big data center operators would like to not discuss any data center details, there is a social and environmental impact the public has the right to review.

The Washington State Department of Ecology has approved Microsoft’s additional permit, but has also scheduled a public hearing in Quincy on Sept. 28 to hear from residents on the topic. The Ecology department conducted an evaluation of the health risks from diesel engine exhaust particulates, and found that the Microsoft expansion, viewed in isolation, is not likely to impact public health.

State officials and Microsoft are required to appear at a public meeting to present and discuss the generator expansion. The Department of Ecology took the opportunity to seek feedback from area residents, citing the growing concentration of data centers.

“Due to the interest expressed by other data companies to expand or build in the Quincy area, Ecology was concerned that the cumulative effect of diesel engine emissions should be assessed,” the state said in announcing the meeting.

Being green means more than your PUE and energy efficiency of equipment.

Here is the presentation referenced.  Gary Palcisko is the presenter.

image

image

image

image  

And note this slide for potential future requirements.

image

Read more