Tue

Feb 13
2007

Tim O'Reilly

Tim O'Reilly

Google Loses Newspaper Copyright Case in Belgium

According to The Globe and Mail:

"A Brussels court ruled in favour of Copiepresse, a copyright protection group representing 18 mostly French-language newspapers that complained the search engine's "cached" links offered free access to archived articles that the papers usually sell on a subscription basis. It ordered Google to remove any articles, photos or links from its sites — including Google News — that it displays without the newspapers' permission. But in the future, it said it would be up to copyright owners to get in touch with Google by e-mail to complain if the site was posting content that belonged to them."

Google says they will appeal, but it seems to me that in this case, the publishers are right. I've gone to the mat for Google (in the context of book search) to defend their fair use right to present snippets (Previous Radar Coverage), but Google needs to find ways to search content behind pay walls. If Google's caching gets around those limits, they should figure out how to work that issue. We know from Google Scholar that they have the means to do so.

As publishers build their own book repositories, Google will need to do the same for book search -- index the content and provide snippets, but serve the actual content through the publisher's site, which may include requirements for payment or other access control. (See Book Search Should Work Like Web Search) It may be that making the content freely available and/or ad supported is a better mechanism, but the content creator should be free to make that choice. Google should provide mechanism, not policy (as Bob Schiefler used to say about the X Window System.)

That being said, the court's judgment can hardly be in the interest of the publishers: "It ordered Google to remove any articles, photos or links from its sites." Once the links disappear, the sites will feel the pain. Most of us get a huge amount of our inbound traffic from Google. The right answer is to respect the paywalls, not to take the stuff out of Google entirely.

This is the problem with court judgments and legislation: it often creates collateral damage.


tags: web 2.0  | comments: 5   | Sphere It
submit:

 
Previous  |  Next

0 TrackBacks

TrackBack URL for this entry: http://blogs.oreilly.com/cgi-bin/mt/mt-t.cgi/5234

Comments: 5

  csven [02.13.07 11:17 AM]

"It may be that making the content freely available and/or ad supported is a better mechanism, but the content creator should be free to make that choice."

Thank you.

  Tom [02.13.07 11:53 AM]

...but Google needs to find ways to search content behind pay walls. If Google's caching gets around those limits, they should figure out how to work that issue.

Google already provides a solution to this problem -- it's called the "noarchive" tag, and it's been standard operating procedure for search vendors for a while now.

Letting the googlebot through your pay wall is also a well-understood technical problem (although Google could probably do more to make it possible to let their bot in while keeping content entirely secure, instead of just mostly secure).

  Danny Sullivan [02.13.07 05:44 PM]

They do have a solution, Tim. It's called the meta noarchive tag, and it's been around for years. If that was really the issue, the publishers could have implemented this. It's explained more at Google here.

It wasn't used -- in the same way robots.txt wasn't used to block pages -- because the publishers weren't interested in using these commonly accepted methods. They were interested in building a case to make Google look like it was somehow just ripping off their content without any controls, to try and force Google to pay them to be included in the Google index.

  Danny Sullivan [02.13.07 05:47 PM]

Sorry, Tom -- missed your comment saying exactly the same thing!

Also, Google can get inside paid walls as it does with Google Scholar and in fact also does for things like Google News Archive and even Google News. At the moment, it does these things on a voluntary basis, working with publishers selectively and not because of paid relationships.

It potentially could do the same for the Belgian publishers, but they've come with an attitude that they are missing out on the "paid" deals they believe Google is doing with others for this type of indexing (which while Google has some paid deals, they are not for the inclusion the Belgian publishers think the deals are for).

  Tim O'Reilly [02.13.07 11:14 PM]

Danny -- I understand about noarchive. But I do think Google could be proactive about moving towards the model where they search behind pay firewalls and support the publishers' charging mechanisms as a way of getting around this kind of situation.

But you're right that I shouldn't be equating the situation with book search too closely to news search.

My point though is that Google should be getting ahead of this issue not just by providing noarchive, but by implementing scholar-like mechanisms for all paid content.

Post A Comment:

 (please be patient, comments may take awhile to post)






Type the characters you see in the picture above.