Previous  |  Next

Mon

Feb 5
2007

Tim O'Reilly

Tim O'Reilly

What We Really Sell to Our Customers is Control

Back in the early days of Red Hat, Bob Young, one of the founders and the first CEO, used to say, "What we really sell to our customers is control." His argument was that one of the benefits of open source software is freedom from vendor lock-in. If you don't like the product, you can get under the hood and change it yourself, or switch to another vendor offering a comparable product based on the same underlying technology.

I had this thought come back around in a Web 2.0 context this past week, with an eerie twist. It was two proprietary Web 2.0 software companies -- enterprise search company FAST and enterprise mapping company Placebase -- that told me, in two independent meetings, that one of the main benefits they sell to customers is control.

How can that be? The common wisdom is that Web 2.0 is all about openness. Of course, anyone who has thought long and hard about why "data is the Intel Inside" is one of my key Web 2.0 maxims understands that successful Web 2.0 companies each control a vast user-contributed (or at least user-enhanced) database that gets progressively more valuable via network effects. And control over that database belongs to the vendor, not to the user. Another way to put it is that every long tail is ultimately attached to a large animal. (I can't remember which wag :-) originally came up with that line.)

FAST, at whose Fastforward user conference I'm speaking later this week, points out that Google claims more than 80 different factors that control search relevancy. But the knobs and levers controlling those factors are largely opaque to users, even corporate users who are private labeling Google's search technology. By contrast, FAST aims to provide a platform under which the customer controls all the knobs and levers, and can add their own relevancy factors as well. So, for example, at Dell or Best Buy, price and availability are additional factors that need to be seamlessly integrated into search. At CareerBuilder, personalization and profiling (e.g. "job seeker" vs. "job searcher") need to be just another factor in the search. (Frankly, that's the same reason we use MarkLogic's xquery search at O'Reilly -- we have control over context and semantics that we can't get from Google, even though they are better at search than we are in so many other ways.)

I heard the same story from Placebase. Why, I asked, do people pay you when it's so easy to put your own data on a map mashup with Google Maps? There were two answers -- first, that there are many additional types of map-data integration than pushpins. Placebase offers a full suite of GIS-style map visualizations, so that the customer can figure out what type of mapping layers most suit the data they want to show. The second answer was a lot like FAST's: we give the customer more control over what they want to do. It's not one-size fits all.

Now, I'm not claiming that Fast or Placebase is somehow more open than Google. I'm simply pointing out that openness occurs at different layers of the stack -- and so does closedness -- and that a proprietary software application might give more control over your own data than a free-as-in-beer Web 2.0 equivalent. What you're actually paying for is that control. My original argument in The Open Source Paradigm Shift (the precursor paper to What is Web 2.0?) was that much as open standards in hardware had moved the locus of proprietary advantage from hardware to software, open source software would move the locus from software to data, via what we now know of as Clayton Christenson's "law of conservation of attractive profits." Companies like Google and Yahoo! have done huge amounts to contribute back to and support open source software, but they jealously guard their own proprietary advantage, which is in data and algorithms and business processes for gathering and managing that data.

And I'm pointing out that a new class of entrepreneurs are finding ways to "open up" the closed applications of Web 2.0. If you're thinking in the old boxes of open source vs. proprietary, you'll miss companies that are surfing new edges between the two.

And what's more, thinking further along the lines of how to give control back to the users, we can ask ourselves what will be the real "open source" answer to proprietary Web 2.0 databases. We've started to hear about "open data", and companies like Wesabe even have open data policies. (Note: I am on the board of Wesabe.) But open data is only part of the new openness we need to explore. What companies like Fast and Placebase do is give users control over the algorithms that are used to manage the data. As Nate Treloar from Fast said during our call, "Relevancy isn't just an algorithm. It's a platform."

And at least so far, companies like Fast and Placebase risk giving their users control over their algorithms by taking a step backwards from another Web 2.0 principle, that of building systems that get better the more people use them. Companies that build their own local data islands may get more control, but they may lose in other ways. In a world driven by network effects, the system that is most open to user contribution will get better faster than one that is more closed, even if the latter system gives more control over algorithms. In my conversations, I urged both Fast and Placebase to think about encouraging their customers to share data with each other, so that each application on their platform benefits from all of the others. Exactly how to do this, given that companies that choose private label search often have proprietary data that they don't want to share with others, is a difficult exercise, but an important one, not be overlooked just because it's hard. Choosing defaults for what data will be made public is probably the single most important architectural decision for any Web 2.0 company.



tags:   | comments: 7   | Sphere It
submit:

 

0 TrackBacks

TrackBack URL for this entry: http://orm3.managed.sonic.net/mt/mt-tb.cgi/6433

Tim O'Reilly   [02.05.07 02:14 PM]

I had written a last paragraph to this posting that went off on a tangent, so I decided to eliminate it from the main story, but keep it as a comment. Here it is...

The risk of limiting the wellspring of user contribution is even greater because many of the companies choosing proprietary enterprise software equivalents to the big public Web 2.0 applications are doing so partly because they are protective of their own data and mindful of their own competitive advantage. And when you have control over algorithms such as search there is always the temptation to tilt results for your short-term business advantage rather than the real needs of the user. And that's a long, slippery slope. (I'm mindful of Jim Allchin's poignant letter to Bill Gates, released recently on the web, "losing our way." [pdf.]) Companies begun in openness and idealism can lose their way so gradually that one day they wake up and don't even know how it happened.

Didier DURAND   [02.05.07 11:06 PM]

Hi,

Google made today a step in the exact
direction that you describe:

they now publish in their Webmaster Central site a
comprehensive list of all links to your site.

It's real step in more transparency / control
over their Pagerank

Check out http://googlewebmastercentral.blogspot.com/2007/02/discover-your-links.html

Rikard Linde   [02.06.07 05:10 AM]

You write:
"In a world driven by network effects, the system that is most open to user contribution will get better faster than one that is more closed, even if the latter system gives more control over algorithms."

To encourage user contribution I'd say cooperation is a more powerful factor than participation, it lets users contribute in more and better ways. And since proprietary data is limited in its ability to help people cooperate my guess is that open data will win this race.

galeal zino   [02.06.07 06:16 AM]

Agree data defaults critical architectural decision - need heuristical analysis for each *vertical* that incorporates all of the factors within the context of the user, usage, goals, etc. I think there are at least two important levels between open and closed, with numerous sub-levels in between:
1. Proprietary data closed but metadata open (including metadata generated by users and usage)
2. Certain data/metadata open within specific communities, but closed outside of those communities. The communities could be defined by simple, manual type borders, or could be defined (automatically) at granular levels, e.g. user x is a member of community y for data z, but is not a member of community y for any other data.
All that said, I think there is increasingly a temptation to expose too many knobs to the user, at the expense of a simple, clean user experience.

Bernard Struyf   [02.07.07 12:40 AM]

If you take a look at the Safari search engine , you will notice that the end-user can play with a slider to control search relevancy, based on sales rank, popularity or statistical elements.
Although most relevancy implementations of search engines are still opaque (even FAST's integration on Dell or Best Buy doesn't really give any control to the end-user), initiatives such as Yahoo Mindset or the Safari search slider seem more "Web 2.0" to me, as they empower the end-user himself.

Tomas Sancio   [02.11.07 05:40 PM]

Makes perfect sense. Even with proprietary software, in which the inner workings of the application are hidden from the user, one can provide control to them by working exclusively with non-proprietary file formats (mainly XML). This is specially useful for small or new software companies, who want to win over wary customers by showing them that the data they produce will be useful as long as there is XSLT to convert the data, even if your company goes down the drain.

Jeff Thurston   [05.12.07 03:58 PM]

Your point is the basis for a divided argument in the international geospatial data community - those who support freely available data and those who don't.

Many would argue that the geographic data in the United States is out-dated and lacking quality for usefulness. In other countries, where such data is paid for, it is of higher quality, more current and has higher potential reuseability.

Having said this, a compromise and fine example of the concept you are speaking about can be seen in those places where coarse scale information of less detail or even older information is given away freely. This enables use and growth around the tools, while preserving the capture of high quality information. High quality data actually then drives further software development.


Post A Comment:

 (please be patient, comments may take awhile to post)




Remember Me?


Subscribe to this Site

Radar RSS feed

RELEASE 2.0 BACK ISSUES

BUSINESS INTELLIGENCE

CURRENT CONFERENCES