Open Data Map movement demonstrates innovation opportunity for Open Sourced Data Center Initiative

Tim Berners- Lee has a 6 minute TED presentation on the year open data went worldwide.

Map and location services are top scenarios for mobile devices.  Google and Microsoft have their maps.  Nokia bought Navteq and MetaCarta.  Apple bought PlaceBase.  With all the companies creating services, volunteers using an open approach to collaborate can beat the proprietary services.

The MercuryNews reports on Open Street Maps.

Volunteers create new digital maps

By Mike Swift

mswift@mercurynews.com

Posted: 04/09/2010 09:08:55 PM PDT

Updated: 04/10/2010 01:36:26 PM PDT

Ron Perez hikes by a waterfall while a portable GPS device records his tracks as... (Jim Gensheimer)

When Brian "Beej" Hall first heard about an audacious volunteer effort to create an Internet map of every street and path in every city and village on the planet, he was hooked. At the time, the nascent effort had only a few American members, and the U.S. map was essentially a digital terra incognita.

Just a few years later, the Berkeley software engineer is editing digital maps so precise they include drinking fountains and benches in the Bay Area parks where he hikes, and the mapping community has swelled to more than 240,000 global members. The effort, OpenStreetMap, is a kind of grass-roots Wikipedia for maps that is transforming how map data is collected, shared and used — from the desktop to smartphones to car navigation.

The reporter makes the observation of how a nonprofit community can change the map business.

But increasingly, the nonprofit community collaboration model behind OpenStreetMap, which shares all the cartographic data in its maps for free, is also changing the business of mapping, just as Wikipedia changed the business of reference. More and more, the accuracy of searches on Google Maps or directions issued by your car's navigational device are based on data collected by volunteers like Hall and other members of OpenStreetMap's do-it-yourself army.

Part of the reason why OpenStreetMap is popular is the fact that the end users are creating the maps.

OpenStreetMap users say that because their data is collected by people who actually live in a place, it is more likely to be accurate.

"It's the people's map," said Paul Jarrett, director of mapping for CloudMade.

If you are interested in the use of OpenStreetMap in Haiti go here.

We chose to tell the story of 'OpenStreetMap - Project Haiti'.
We all followed the crisis that unfolded following the Haiti earthquake, many of us chose to donate money, a few were flown out and deployed as part of the relief effort. But what practical impact can many have without being there in Haiti itself? Well, during this crisis a remarkable story unfolded; of how people around the world could virtually collaborate and contribute to the on-the-ground operations.

OpenStreetMap - Project Haiti 1

With the little existing physical, political and social infrastructure  now destroyed or damaged, the situation was especially challenging for aid agencies arriving on the ground. Where are the areas most in need of assistance, how do we get there, where are people trapped under buildings, which roads are blocked? This information is important to the rescue agencies immediately after the event, and to the longer rebuilding process. In many developing countries, there is a lack of good mapping data and particularly after a crisis, when up-to-date information is critical to managing events as they evolve.
Enter OpenStreetMap, the wiki map of the world, CrisisMappers and an impromptu community of volunteers who collaborated to produce the most authoritative map of Haiti in existence. Within hours of the event people were adding detail to the map, but on January 14th high resolution sattelite imagery of Haiti was made freely available and the Crisis Mapping community were able to trace roads, damaged buildings, and enter camps of displaced people into OpenStreetMap. This is the story of OpenStreetMap - Project Haiti:

There are many who think the Open Source Data Center Initiative will not work.  There are a lot of people who thought OpenStreetMaps wouldn't work too.

Read more

Next Advisor for GreenM3 NPO, Peter Horan, pushing the edge of the network to be close to customers

Our first industry advisor was Mike Manos, our next is Peter Horan. Peter is unknown to most of the data center audience as he is an executive who has worked on the edge of innovation, not in the hub of data center activity.  Peter does have data center experience as the Sr. VP executive for InterActive Media at the time of ask.com's data center construction at Moses Lake, WA.  Chuck Geiger was CTO of ask.com at the time, and stated.

“Moses Lake is an ideal location due to its cooperative business environment, access to low cost, renewable power and superior network connectivity,” said Chuck Geiger, Chief Technology Officer of Ask.com. “With these inherent benefits, Eastern Washington is the right choice for Ask.com as we expand our computing infrastructure to support our growth and expanded search services.”

Peter has had the executive's view of building a large data center, yet he has some very innovative, forward thinking ideas and a powerful network.  Which brings up a presentation that Peter made discussing the "Edge of the Network."

image

I've known Peter for many years, including his time as Sr. VP/Publisher of ComputerWorld, CEO of DEVX.com, about.com, allbusiness.com, and was an obvious candidate for the GreenM3 NPO.

image

Here is a video where Peter presents the ideas to get closer to customers.  In the same way Peter encourages the audience to get close to customers, the goal of GreenM3 is to build a closer connection to customers, using open source techniques.

image

A person who we want to talk to in Peter's network is Chuck Geiger.

Chuck Geiger
Partner - Technology

Chuck has significant experience running some of the largest online transaction product organizations and most visited sites in the world, including as CTO of Ask.com, CTO of PayPal, VP Architecture of eBay, and executive positions at InterActive Corp., Gateway and Travelocity.


At InterActive Corp, Chuck was responsible for managing a consolidated data center strategy for IAC portfolio companies including Ask.com, Evite.com, CitySearch.com, Ticketmaster, and Match.com. Chuck also was responsible for the technology organization at Ask.com including Engineering, Architecture, Program Management, QA, IT, and Operations.


At PayPal, Chuck was responsible for the product development organization which includes Product Management, Design, Engineering, Architecture, Operations, IT, QA, Project Management, Content, and Localization, running a team of approximately 550 professionals.
At eBay, Chuck was responsible for the migration to the new generation system architecture and platform.

BTW, Peter's day job is Chairman of Goodmail.

About Goodmail Systems

Goodmail Systems is the creator of CertifiedEmail™, the industry’s standard class of email. CertifiedEmail provides a safe and reliable means for consumers to easily identify authentic email messages from legitimate commercial and nonprofit email senders. Each CertifiedEmail is sent with a cryptographically secure token that assures authenticity and is marked in the inbox with a unique blue ribbon envelope icon, enabling consumers to visually distinguish email messages which are real and sent from email senders with whom they have a pre-existing relationship.

We welcome Peter's passion for technical innovation and the environment.

Read more

Long Now, Long View, Long Lived Data Center, a 10,000 year clock - a 10,000 year data center?

I am currently thinking of rules for the ontology in data center designs.  Translated, I am trying to figure out the principles, components, and relationships for the Open Source Data Center Initiative. 

This is a complex topic to try and explain, but I found an interesting project the Long Now started by a bunch of really smart people, Jeff Bezos, Esther Dyson, Mitch Kapor, Peter Schwartz, and Steward Brand.  Here is a video discussing the idea of a 10,000 year clock.

 

But, what I found interesting was their long term approach and transparency that we will be using in the Open Source Data Center Initiative.  And, now thinking a Long View is part of what we have as principles.

Here are the principles of the Long Now Clock that make a lot of sense to use data center design.


These are the principles that Danny Hillis used in the initial stages of designing a 10,000 Year Clock. We have found these are generally good principles for designing anything to last a long time.

Longevity

With occasional maintenance, the clock should reasonably be expected to display the correct time for the next 10,000 years.

Maintainability

The clock should be maintainable with bronze-age technology.

Transparency
It should be possible to determine operational principles of the clock by close inspection.
Evolvability
It should be possible to improve the clock with time.
Scalability
It should be possible to build working models of the clock from table-top to monumental size using the same design.

Some rules that follow from the design principles:

Longevity:
Go slow
Avoid sliding friction (gears)
Avoid ticking
Stay clean
Stay dry
Expect bad weather
Expect earthquakes
Expect non-malicious human interaction
Dont tempt thieves
Maintainability and transparency:
Use familiar materials
Allow inspection
Rehearse motions
Make it easy to build spare parts
Expect restarts
Include the manual
Scalability and Evolvabilty:
Make all parts similar size
Separate functions
Provide simple interfaces

Why think about a 10,000 year clock, because thinking about slowness teaches us things we don't have time when think only of speed.

Hurry Up and Wait

The Slow Issue > Jennifer Leonard on January 5, 2010 at 6:30 am PST

018-futurists-1

We asked some of the world’s most prominent futurists to explain why slowness might be as important to the future as speed.

Julian Bleecker
Julian Bleecker, a designer, technologist, and co-founder of the Near Future Laboratory, devises “design-to-think experiments” that focus on interactions away from conventional computer settings. “When sitting at a screen and keyboard, everything is tuned to be as fast as possible,” he says. “It’s about diminishing time to nothing.”
So he asks, “Can we make design where time is inescapable and not be brought to zero? Would it be interesting if time were stretched, or had weight?” To test this idea, Bleeker built a Slow Messaging Device, which automatically delayed electronic (as in, e-mail) messages. Especially meaningful messages took an especially long time to arrive.

Read more: http://www.good.is/post/hurry-up-and-wait#ixzz0jmOEcLg4

The biggest unknown problems in data centers are those things that we didn't think were going to happen in the future.  And, this leaves the door open to over-engineering, increasing cost, brittleness of systems, and delays.  Taking a Long View what will the future possibly look like can help you see things you normally wouldn't.

image

Read more

Sustainable Farming Method applied to Sustainable Data Centers, Dan Barber's entertaining how I fell in love with a fish, it's about relationships

This is a video that has 5 stars. It is entertaining, funny and educational.

Here is Huffington post article about the video.

Dan Barber: How I Fell in Love With a Fish

Chef Dan Barber squares off with a dilemma facing many chefs today: how to keep fish on the menu. With impeccable research and deadpan humor, he chronicles his pursuit of a sustainable fish he could love, and the foodie's honeymoon he's enjoyed since discovering an outrageously delicious fish raised using a revolutionary farming method in Spain.

Here is a picture of the fish farm.

image

Which is different than a typical fish farm.

image

One of the best lines that gives you the method, the secret to success in the sustainable farm is the biologist isn't an expert in fish, but is an expert in relationships.  If you go to time mark 8:10 you can listen to this.

Dan asks How did he become an expert on fish?

Fish??  I don't know anything about fish.  I am an expert on relationships.

Why is this so important?  Because the people who are doing the most innovative data center work understand the relationships of the site to the building to the IT equipment to the software and the services provided.  This is what good system engineers know in mature industries.  This is beyond the data center building with its power and cooling systems.  The enlightened are looking at the energy supply chain with a focus on cost, carbon impact, and changes in the supply chain in the future.  What is the future of services and applications that need to run on servers, storage, and networks.  This is one damn hard problem to address as the silos in data center are powerful and entrenched in a companies organization and the rest of the industry.

I just came back from Missouri and got a chance to talk in more detail to the Civil Engineering company, Allstate Consulting who is working on site analysis of the Ewing Industrial Park.  Over the past 9 months there are a variety of people who are being exposed to data centers who had no previous data center experience.  Yet, there are many instances of where potential site users are surprised on the engineering analysis and drawings prepared.

Having a pizza at Shakespare's Pizza in Columbia, MO, with Chad Sayre, VP of Design Services at Allstate Consulting.

Chad W. Sayre, PE, Vice-President

Mr. Sayre obtained a Master of Science and a Bachelor of Science in Civil Engineering from University of Missouri-Columbia.  Since 1994, Mr. Sayre has been a Project Engineer for Allstate Consultants.  In 2000, he became Principal/Vice President and Chief of Design Services for the firm.  He is responsible for municipal and land development engineering projects including forensic, water, wastewater, land planning, environmental compliance permitting, highway design, hydraulics, and stormwater projects.  He is closely involved with construction administration, inspection, specification preparation of public and private projects and has a considerable amount of experience in expert testimony.

He was telling the story of how he attended DataCenterDynamics conference and was in the bar (which is common networking method at all DataCenterDynamics event), and was telling a big data center customer, "A year ago, I didn't know shit about data centers." 

But, I'll tell you what Chad does know is the methane gas production issues from the adjacent land fill.  How the topography of the site can be used to create isolation areas to protect the site and change building design. How easy it is to dig trenches to connect to additional fiber. How BioMass can be used to generate renewable energy. How construction techniques that have been applied to multiple other industries can be used in manner similar to what Microsoft has proposed.

These facilities will not be pretty and might actually resemble the barns I spent so much time around during my childhood in rural Illinois. That, combined with the fact that these facilities will be substantially lower cost per megawatt to build and substantially lower cost to run, makes it very easy to become excited about what we’re doing here.

William Gibson said it best: “The future is here, it’s just not widely distributed yet.”

When you look for who knows how to build sustainable, green, low carbon data centers, look for those who understand relationships.  I am tired of hearing people tout their latest hardware as the answer to the problem, but they can't explain how this new equipment effects the rest of the system.

It would be pretty damn funny, but almost impossible to create a video like Dan Barber did for "How I fell in love with a Data Center"

Actually, I do know lots of people who fell in love with their data centers.  But many times it is more like this image, high density, low PUE, with little discussed on the carbon impact or waste.

image

Wouldn't it be cool if data centers could help the environment?

image

Read more

Google Warehouse Scale Computing pattern harvested, solving the current or future performance problems

The Open Source Data Center Initiative is using a Pattern based approach.

In software engineering, a design pattern is a general reusable solution to a commonly occurring problem in software design. A design pattern is not a finished design that can be transformed directly into code. It is a description or template for how to solve a problem that can be used in many different situations.

I was reading Google's Warehouse Scale Computing document which can be daunting with its 120 pages of dense topics.  One of the points made which is an example of a design pattern is under the following conditions.

Key pieces of Google’s services have release cycles on the order of a couple of weeks compared to months or years for desktop software products. Google’s front-end Web server binaries, for example, are released on a weekly cycle, with nearly a thousand independent code changes checked in by hundreds of developers— the core of Google’s search services has been reimplemented nearly from scratch every 2 to 3
years.

This may not sound like your environment, but it is common in agile dynamic SW development at Google, start-ups and other leading edge IT shops.

Agile methods generally promote a disciplined project management process that encourages frequent inspection and adaptation, a leadership philosophy that encourages teamwork, self-organization and accountability, a set of engineering best practices intended to allow for rapid delivery of high-quality software, and a business approach that aligns development with customer needs and company goals.

The old way of purchasing IT hardware to support an application's SLA is a lower priority.  The new way is to add hardware capabilities to support the rapid innovation in SW development.

A beneficial side effect of this aggressive software deployment environment is that hardware architects are not necessarily burdened with having to provide good performance for immutable pieces of code. Instead, architects can consider the possibility of significant software rewrites to take advantage of new hardware capabilities or devices.

BTW, immutable in SW means which applies to many legacy systems.

In object-oriented and functional programming, an immutable object is an object whose state cannot be modified after it is created. This is in contrast to a mutable object, which can be modified after it is created.

Problem: How to improve the performance per watt  in IT efficiencies with data center infrastructure and hardware?

Options:

  1. Improve data center efficiency, aka PUE.
  2. Buy more efficiency IT HW.
  3. Improve HW utilization with virtualization and server consolidation.
  4. Add new hardware capabilities that support the future of software.

Solution: even though 1 - 3 are typical, the efficiencies from #4 could be sizeable larger.  Some part of the data center and IT hardware should be designed for the future applications vs. making the future applications run on what the past applications require.

Examples of technologies are NVidia's GPU, solid state memory, startups with new hardware designs like www.tilera.com, and complete re-architecture of the data center system.

People are working on the complete re-architecture of the data center system as the performance per watt gains are huge.

How many data centers are designed for the current hardware vs the future? 50%, 75%, 90%, 95%, 98%

Should data centers be designed for a 5 year lifespan vs. 20 - 30 to support more rapid innovation?  Then, be upgradable?

Read more