Two different ways to implement geo-redundancy data storage - Infineta and EMC VPlex

James Hamilton discusses inter-datacenter replication and geo-redundancy which was quite easy for me to get my mind wrapped around as the issues discussed have a lot in common to work I did on Microsoft's Branch Office Infrastructure Solution (BOIS) and the issues with WAN.

Inter-Datacenter Replication & Geo-Redundancy

Wide area network costs and bandwidth shortage are the single most common reason why many enterprise applications run in a single data center. Single data center failure modes are common. There are many external threats to single data center deployments including utility power loss, tornado strikes, facility fire, network connectivity loss, earthquake, break in, and many others I’ve not yet been “lucky” enough to have seen. And, inside a single facility, there are simply too many ways to shoot one’s own foot. All it takes is one well intentioned networking engineer to black hole the entire facilities networking traffic. Even very high quality power distribution systems can have redundant paths taken out by fires in central switch gear or cascading failure modes. And, even with very highly redundant systems, if the redundant paths aren’t tested often, they won’t work. Even with incredibly redundancy, just having the redundant components in the same room, means that a catastrophic failure of one system, could possibly eliminate the second. It’s very hard to engineer redundancy with high independence and physical separate of all components in a single datacenter.

With incredible redundancy, comes incredible cost. Even with incredible costs, failure modes remain that can eliminate the facility entirely. The only cost effective solution is to run redundantly across multiple data centers. Redundancy without physical separation is not sufficient and making a single facility bullet proof has expenses asymptotically heading towards infinity with only tiny increases in availability as the expense goes up. The only way to get the next nine is have redundancy between two data centers. This approach is both more available and considerably more cost effective.

The solution James references is from Infineta.

Last week I ran across a company targeting latency sensitive cross-datacenter replication traffic. Infineta Systemsannounced this morning a solution targeting this problem: Infineta Unveils Breakthrough Acceleration Technology for Enterprise Data Centers. The Infineta Velocity engine is a dedupe appliance that operates at 10Gbps line rate with latencies under 100 microseconds per network packet. Their solution aims to get the bulk of the advantages of the systems I described above at much lower overhead and latency. They achieve their speed-up three ways: 1) hardware implementation based upon FPGA, 2) fixed-sized, full packet block size, 3) bounded index exploiting locality, and 4) heuristic signatures.

Infineta provides a technology overview

Technology Overview

Infineta Systems delivers solutions based on several technologies that significantly reduce the amount of traffic running across today’s data center WAN interconnect. Our groundbreaking innovation centers on the patent-pending Velocity Dedupe Engine™, the industry’s first-ever hardware deduplication (“dedupe”) engine. Unlike alternatives, the Velocity Dedupe Engine enables our solutions to maintain the highest levels of data reduction at multi-gigabit speeds while guaranteeing port to port latencies in the few 10s of microseconds. As a result, Infineta’s solutions enable customers to accelerate all data center applications (such as replication and backup) – including ones that are highly latency sensitive. It does so while reducing overall costs incurred by this growing, bandwidth-hungry traffic.

Technology Architecture

Distributed System Architecture

Unlike traditional acceleration solutions that are assembled around monolithic processing environments, Infineta’s solutions are designed from the ground-up around a distributed processing framework. Each core feature set is implemented in a dedicated hardware complex, and they are all fused together with high-speed fabric, guaranteeing wire speed acceleration for business-critical traffic.

Hardware-based Data Reduction

Data reduction is carried out purely in hardware in a pipelined manner, allowing the system to reduce enterprise WAN traffic by as much as 80-90 percent.

Fabric-based Switching

Infineta’s solutions are built on a massive scale, fully non-blocking switch fabric that can make precise switching decisions in the face of sustained traffic bursts.

Multi-gig Transport Optimization

Infineta’s solutions employ multi-core network processors to carry out transport-level acceleration at multi-gigabit speeds. By making key resource decisions at wire-speed, the system is able to maintain up to 10Gbps traffic throughput while working around detrimental WAN characteristics, such as packet loss.

And EMC has VPlex for Virtualized Storage across multiple data centers.

image

News Summary:

  • EMC advances Virtual Storage with industry’s first distributed storage federation capabilities, eliminating boundaries of physical storage and allowing information resources to be transparently pooled and shared over distance for new levels of efficiency, control and choice.
  • Groundbreaking EMC VPLEX Local and VPLEX Metro products address fundamental challenges of rapidly relocating applications and large amounts of information on-demand within and across data centers which is a key enabler of the private cloud.
  • Future EMC VPLEX versions will add cross-continental and global capabilities, allowing multiple data centers to be pooled over extended distances, providing dramatic new distributed compute and service-provider models and private clouds.
Read more

With Cheap Plentiful Natural Gas, Should Data Centers be sited near Natural Gas supply for lower cost and carbon footprint?

WSJ has an article on Shale Gas.

Shale Gas Will Rock the World

Huge discoveries of natural gas promise to shake up the energy markets and geopolitics. And that's just for starters.

By AMY MYERS JAFFE

There's an energy revolution brewing right under our feet.

Over the past decade, a wave of drilling around the world has uncovered giant supplies of natural gas in shale rock. By some estimates, there's 1,000 trillion cubic feet recoverable in North America alone—enough to supply the nation's natural-gas needs for the next 45 years. Europe may have nearly 200 trillion cubic feet of its own.

The author is convinced this will change the energy industry which data centers are a part of.

I have been studying the energy markets for 30 years, and I am convinced that shale gas will revolutionize the industry—and change the world—in the coming decades. It will prevent the rise of any new cartels. It will alter geopolitics. And it will slow the transition to renewable energy.

The author touches on thoughts on renewable energy.

Then, I think we still need to invest in renewables—but smartly. States with renewable-energy potential, such as windy Texas or sunny California, should keep their mandates that a fixed percentage of electricity must be generated by alternative sources. That will give companies incentives and opportunities to bring renewables to market and lower costs over time through experience and innovation. Yes, renewables may seem relatively more expensive in those states as shale gas hits the market. And, yes, that may mean getting more help from government subsidies. But I don't think the cost would be prohibitive, and the long-term benefits are worth it.

Still, I don't believe we should set national mandates—which would get prohibitively expensive in states without abundant renewable resources. Instead of pouring money into subsidies to make such a plan work, the federal government should invest in R&D to make renewables competitive down the road without big subsidies.

Is natural gas supply a new criteria for data center sites running cogeneration plants or fuel cells?

Read more

Not Calculating PUE ROI, 1 of 9 Top Data Center Mistakes

Lee Technologies sent over their Top 9 data center mistakes paper, and I liked it right from the start as they made the point people don't calculate an ROI for PUE performance.  PUE is closest to telling an efficiency of a data center which to a layman is the closest we have for data centers for a MPG, but who specifies a high MPG number and doesn't think about how much it costs for the extra MPG above the norm.

image

The Top 9 Mistakes in Data Center Planning: The Total Cost of Ownership Approach to Building or Expanding Data Centers

Why do so many data center builds and expansions fail? This white paper answers the question by revealing the top 9 mistakes organizations make when designing and building new data center space, and examines an effective way to achieve success through the Total Cost of Ownership (TCO) approach.

I spend too much time writing for other people, and I could go on how I would rewrite the paper, but overall an excellent job bringing up issues few think about.

The top three points that I think would get people's attention are ironically the last 3 of 9 tips.

# 7

Big Mistake #7:

Misunderstanding PUE

Many times, organizations set a PUE goal with all the proper intentions
but the calculation does not take into account all factors that should be
considered.


You need to fully understand what the ROI is on capital expenses to
reach your goals. You need to ask yourself, what is the TC O relative to
the target PUE?

#8 has three parts, and I  like this one best as this hidden cost of LEED is why builders and consultants push LEED.  People think they are getting LEED for free, because they are not counting the cost of certification.

Big Mistake #8:
Misunderstanding LEED certification
To date, the U.S. Green Building Council (USGBC) has not set specific
criteria for data center LEED criteria.

There will be costs related to receiving certification. Failure to take
these related expenses into account will impact your TCO and
business decision planning processes.

And #9, an overall problem with data center design.

Big Mistake #9:
Overcomplicated designs
As stated earlier, simple is better. Regardless of the target tier rating
you have chosen, there are dozens of ways to design an effective
system. Too often, redundancy goals drive too much complexity. Add
in the multiple approaches to building a modular system and things get
complicated fast.

Read more

Triple Bottom Line - social, economic, and environment responsibility, can BP be Green in light of the Gulf Spill?

Being green is an overused term.  Many of the enlightened speak of triple bottom line.

The triple bottom line (abbreviated as "TBL" or "3BL", and also known as "people, planet, profit" or "the three pillars"[1]) captures an expanded spectrum of values and criteria for measuring organizational (and societal) success: economic, ecological and social. With the ratification of the United Nations and ICLEI TBL standard for urban and community accounting in early 2007, this became the dominant approach to public sector full cost accounting. Similar UN standards apply to natural capital and human capital measurement to assist in measurements required by TBL, e.g. the ecoBudget standard for reporting ecological footprint.

BP is claiming its Green.

BP touts itself as 'green,' but faces PR disaster with 'BP oil spill'

GALLERY

Deepwater Horizon explosion in Gulf of Mexico sends oil slick toward U.S. coastline

Cleanup and containment efforts continue after an oil platform explosion in the Gulf of Mexico. Oil is leaking into the water at a rate of up to 5,000 barrels a day.

By Paul Farhi

Washington Post Staff Writer
Thursday, May 6, 2010

What do you call a gigantic man-made disaster that is threatening to despoil the ecosystems and wreck the economies of the Gulf Coast? The answer is important, if you happen to be one of the companies responsible for it.

The massive slick spreading toward Louisiana has gone by several names since crude oil began gushing from a damaged drilling rig on April 20. Media accounts have referred to it as "the Gulf oil spill," "the Deepwater Horizon spill" and the "Gulf Coast disaster."

President Obama, leaving little doubt about whom he considers responsible for the epic mess, put a brand name on it in remarks in Louisiana on Sunday. The president dubbed it "the BP oil spill," after the company (formerly British Petroleum) that leased the now-damaged drilling platform. The Environmental Protection Agency refers to it the same way in its official pronouncements.

Read more

The Greenest Data Center may be the one you don't build

A regular question is what is a green data center.  One simple answer could be I was able to avoid building a new one.  Zero carbon footprint is easy to achieve with no building.

TriplePundit discusses Dell, HP, and Wells Fargo efforts to not build a data center.

IT Giants: The Greenest Data Center is the One That Isn’t Built

By Sheila Samuelson | May 10th, 2010 View Comments

Image source: Data Center Knowledge

Last month, Dell made the somewhat shocking announcement that it may never build another data center. The company was referring to the fact that it’s doubled its workload using no extra power and building no new data centers, simply by squeezing more capacity out of its existing servers. With an industry standard for data server utilization at about 12 to18 percent, there is ample room for improvement. What Dell realized was that by getting rid of its underutilized assets and swapping out the oldest and most outdated 25 percent of servers each year for the newest virtualization models, it would easily recoup its capital expenditures through reduced energy costs.

Read more