Frequent thinker, occasional writer, constant smart-arse

Category: Open standards

A billion dollar opportunity with video

When Google made an offer for On2, I was dumbfounded. I wrote to a friend working at Google the following:

Phat. But I’m confused. How does Google benefit by making the codec free? I understand Google’s open culture, but for 100million, really? They help the world, but what’s the incentive for Google? (Other than of course, controlling it).

The reply: “incentive = greater adoption of HTML 5 = apps are written for HTML 5 = apps can be monetized using Adsense”.

Interesting perspective from a smart Googler who had no real insider information. But no cigar.

Newsteevee posted a follow up article today on what Google is going to do with this technology, quoting the Free Software Foundation. What really made me get thinking was this (emphasis mine:

Google’s Open Source Programs Manager Chris DiBona had previously argued that Ogg Theora would need codec quality and encoding efficiency improvements before a site as big as YouTube could use it as its default video codec. The FSF now writes in its letter that it never agreed with these positions, but that Google must have faith in VP8 being a better codec if it invested its money in it (Google spent a total of about $133 million on ON2).

The open source advocacy group apparently realized that Google wouldn’t switch codecs from one day to another, which is why it suggests a number of smaller steps to make VP8 mainstream. “You could interest users with HD videos in free formats, for example, or aggressively invite users to upgrade their browsers (instead of upgrading Flash),” the letter reads, adding that this would eventually lead to users not bothering to install Flash on their computers.

Think about that for a second: video on the web finally becomes free for real and open, becoming a core infrastructure to the online world – but the default is crappy. Don’t like crappy? Well Mr and Ms consumer, if you want High Definition, you need to pay for a subscription to a premium codec by the already dominate Adobe or another rising star. Assuming you get the whole word watching video and only 1% convert – holy crap, isn’t that a brilliant business model?

Bono, the lead singer of the band U2 wrote in an op-ed piece in the New York Times the following recently:

The only thing protecting the movie and TV industries from the fate that has befallen music and indeed the newspaper business is the size of the files”

Simple but profound insight from the famed entertainer. So with this fairly obvious logic, why isn’t the movie industry (backed by Google and Apple) innovating business models in this area? Value comes from scarcity – and quality is the best way of doing it. The reason why box office sales and Blu-ray broke a record in 2009, is because the quality is worth the premium for consumers.

What’s the incentive for Google, to answer my own question? The return on investment to be associated with a default open technology that you give the option to upgrade to users, is a billion dollar business waiting to happen. Doing no evil to the world and securing future growth at the same time sounds like a Google business in the making,

Why open wins

Open standards matter, but so does the water; and just like water is not what creates a Mona Lisa or a Hoover Dam alone, so too do open standards not really matter that much to what we are trying to do with the DataPortability Project in the longer term. But they matter for the industry, which is why we advocate for them. Here’s why.

Hoover dam

Bill Washburn is one of the soft-spoken individuals that has driven a lot of change, like leading the charge to open government technology (the Internet as we know it) to the rest of the world. He’s been around long enough to see trends, so I asked him: why does open always win? What is it about the walled garden that makes it only temporary?

Bill gave me two reasons: technologies need to be easy to implement and they also need to be cheap. It may sound obvious, but below I offer my interpretation why in the context of standards

1) Easy to implement
If you are a developer constantly implementing a standard, you want the easiest one to implement. Having to learn a new standard each time you need to do something is a burden – you want to learn how to do something once and that’s it. And if there is a choice to implement two standards that do the same thing, guess which one will win?

That’s why you will see the technically inferior RSS dominate over ATOM. Both allow syndication and give the end-user the same experience, but for a developer trying to parse it, ATOM is an absolute pain in the buttocks. Compare also JSON and XML – the former being a data structure that’s not even really a standard, and the latter which is one of the older data format standards on Internet. JSON wins out for using asynchronous technologies in the web2.0 world, because it’s just easier to do. Grassroots driven micro-formats and W3C endorsed RDF? Same deal. RDF academically is brilliant – but academic isn’t real world.

2) Cheap to implement
This is fairly obvious – imagine if you had two ways of performing something that did the same thing, but one was free and the other had licensing costs – what do you think a developer or company will use? Companies don’t want to pay licensing fees, especially for non-core activities; and developers can’t afford license fees for a new technology. Entities will bias their choices to the cheaper of the two, like free.

I think an interesting observation can be made about developer communities. Look at people that are the .Net community, compared to say something like Python advocates. You tend to find Python people are more open to collaboration, meetups, and other idea exchanges rather than the .Net developers who keep to themselves (a proprietary language). With the Microsoft owned .Net suite requiring a lot more costs to implement, it actually holds back the adoption of the technology to dominate the market. If people aren’t collaborating as much when compared to rival technologies, that means less innovation, more costs to learning – a longer term barrier to market adoption.

The most important point to make is on the actual companies that push these standards. Let’s say you are Facebook pushing your own standard, which although free, could only be modified by and adapted by the Facebook team. That’s going to cost resources – at the very least, a developer overseeing it. Maybe a team of evangelists to promote your way of thinking; a supervisor to manage this team. If you are the sole organisation in charge of something, it’s going to cost you (not anyone else) a lot of money.

Bridge being built on the Hoover dam

Compare that to an open community effort, where lots of companies and people pool their resources. Instead of one entity bearing the cost, it’s hundreds of entities bearing the cost. On a singular basis, it’s actually cheaper to create a community driven standard. And honestly, when you think about it, why a company fights over what standard gets implemented has nothing to do with their core strategic objectives. Sure they might get some marketing out of it (as the Wikipedia page says “this company created this standard”), but realistically, it’s rewarding more the individuals within these companies who can now put on their resume “I created this technology that everyone is using now”.

Why Open wins
In the short run, open doesn’t win because it’s a longer process, that in part relies on an industry reacting to a proprietary approach. In the long run, Internet history has proven that the above two factors always come to dominate. Why? Because infrastructure is expensive to build and maintain, and usually, it’s better to pool our efforts to build that infrastructure. You don’t want to spend your money on something that’s for the public benefit, only to have no one in the public using it – do you, Mr Corporate Vice-President?

Data portability allows mashup for Australian bush fire crisis

Last night in Australia, one of the states developed a series of bush fires that have ravaged communities – survivors describe it as “raining fire” that came out of no where. As I write this, up to 76 people have been killed.

Victorian AU Fires 2009
The sky is said by Dave Hollis to look how it is in the movie ‘Independence Day’

An important lesson has come out out of this. First, the good stuff.

Googler Pamela Fox has created an invaluable tool to display the bush fires in real time. Using Google technologies like App engine and the Maps API (which she is the support engineer for), she’s been able to create a mashup that helps the public.

She can do so because the Victorian Fire department supports the open standard RSS. There are fires in my state of New South Wales as well, but like other Fire Department’s in Australia, there is no RSS feed to pull the data from (which is why you won’t see any data on the map from there) It appears states like NSW do support RSS for updates, but it would be more useful if there was some consistency – refer to discussion below about the standards.

For further information, you can read the Google blog post.

While the Fire Department’s RSS allows the portability of the data, it doesn’t have geocodes or a clear licence for use. That may not sound like a big deal, but the ability to contextualise a piece of information in this case matters a hell of a lot.

As a workaround, Pamela sent addresses through the Google geocoder to develop a database of addresses with latitude and longtitude.

GeoRSS and KML
In the geo standards world, two dominant standards exist that enable the portability of data. One is an extension to RSS (GeoRSS) that allows you to extend an RSS feed to show geodata. The other in Keyhole Markup Language, which was a standard developed by Google. GeoRSS is simply modifying RSS feeds to be more useful, while KML is more like how HTML is.

If the CFA and any other websites had supported them either of these standards, it would have made life a lot more easier. Pamela has access to Google resources to translate the information into a geocode and even she had trouble. (Geocoding the location data was the most time-consuming of the map-making process.)

The lessons
1) If you output data, output it in some standard structured format (like RSS, KML, etc).
2) If you want that data to be useful for visualisation, include both time and geographic (latitude/longitude information). Otherwise you’re hindering the public’s ability to use it.
3) Let the public use your data. The Google team spent some time to ensure they were not violating anything by using this data. Websites should be clearer about their rights of usage to enable mashers to work without fear
4) Extend the standards. It would have helped a lot of the CFA site extended their RSS with some custom elements (in their own namespace), for the structured data about the fires. Like for example <cfa:State>Get the hell out of here</cfa>.
5) Having all the Fire Department’s using the same standards would have make a world of difference – build the mashup using one method and it can be immediately useful for future uses.

Pamela tells me that this is the fifth natural disaster she’s dealt with. Every time there’s been an issue of where to get the data and how to syndicate it. Data portability matters most for natural disasters- people don’t have time to deal with scraping HTML (didn’t we learn this with Katrina?).

Let’s be prepared for the next time an unpredictable crisis like this occurs.