News Publishers Want To Change Robots.txt; Want To Make Sure Their Content Is Less Useful

from the deep-misunderstandings dept

Following on the speech given earlier this month by the head of the Associated Press, where it was made clear that the AP and news organizations still think that they can be gatekeepers of news, a bunch of publishers along with the AP are now trying to revise robots.txt so that they can hide content on a more selective level. Now, it is true that robots.txt can be rather broad in its sweep. But it's rather telling that it's the publishers who banded together and are telling search engines what changes are needed, rather than working with the search engines to come up with a reasonable solution. In the meantime, there really are some simple solutions if you don't want content indexed by search engines -- but we've yet to fully understand why publishers are so upset that Google, Yahoo and others are sending them so much traffic in the first place.
Hide this

Thank you for reading this Techdirt post. With so many things competing for everyone’s attention these days, we really appreciate you giving us your time. We work hard every day to put quality content out there for our community.

Techdirt is one of the few remaining truly independent media outlets. We do not have a giant corporation behind us, and we rely heavily on our community to support us, in an age when advertisers are increasingly uninterested in sponsoring small, independent sites — especially a site like ours that is unwilling to pull punches in its reporting and analysis.

While other websites have resorted to paywalls, registration requirements, and increasingly annoying/intrusive advertising, we have always kept Techdirt open and available to anyone. But in order to continue doing so, we need your support. We offer a variety of ways for our readers to support us, from direct donations to special subscriptions and cool merchandise — and every little bit helps. Thank you.

–The Techdirt Team

Filed Under: publishers, robots.txt
Companies: associated press


Reader Comments

Subscribe: RSS

View by: Time | Thread


  1. identicon
    JS Beckerist, 30 Nov 2007 @ 10:25am

    Simple solutions:

    A: Put the content you don't want crawled on a different server and blacklist all known bots from connecting to that server.
    B: Use a login script.

    link to this | view in thread ]

  2. identicon
    Johan, 30 Nov 2007 @ 11:12am

    I've said it before...

    and I'll say it again - Google, Yahoo, and the others need to completely remove news publishers that complain from their listings. Poof! Gone! Problem solved!

    Unfortunately for the publishers no more listings equals no more traffic. Bye-bye AP, bye-bye archaic newspapers that can't make the move into the 21st century.

    If the search engines would do this the news publishers would quickly see the value of the listings.

    link to this | view in thread ]

  3. identicon
    data64, 30 Nov 2007 @ 11:18am

    More info on proposed changes ?

    Wish there was more information on the proposed changes. The Acap website is also not very clear. There is some information on their reasoning in their FAQs.

    update:
    A pdf document with the proposed changes.

    link to this | view in thread ]

  4. identicon
    Carme, 30 Nov 2007 @ 12:41pm

    Why it does make sense

    The fact that there are no business models that make use of this fine-grained control now doesn't mean there won't be such in the future. They are smartly making sure the technology is already in place so they are free to experiment with such business models, and if they find a successful one stick with it. While it's possible they won't find a successful model, they certainly won't find it without the technology to back it up.

    You may believe that no such model exists, but bear in mind that a failed experiment is as important as a successful one. For example, if the music industry started off selling unprotected MP3, they would probably get scared of the inevitable piracy, however minor, that would ensue and fold out of the digital media market. Now that they tried various DRM schemes - and failed miserably - they are much more committed to selling unprotected digital media and much more pragmatic as to what is an acceptable level of piracy. Similarly, while such an experimentation period may not help the publishing industry actually find a better Internet strategy, it will certainly make sure they are more informed and committed when they do pick their strategy.

    And anyway, having more control can never be bad. Since the courts seem to agree the publishers do have a say as to how their content is used, they might as well use that power to put in place the technology to enforce control. Then they can decide to just continue with the status quo, and change it when they find it beneficial.

    link to this | view in thread ]

  5. identicon
    T, 3 Dec 2007 @ 9:45am

    Be patient. ACAP is a proposal and voluntary. Those that that would need to use it will decide if it has value.

    link to this | view in thread ]


Follow Techdirt
Essential Reading
Techdirt Deals
Report this ad  |  Hide Techdirt ads
Techdirt Insider Discord

The latest chatter on the Techdirt Insider Discord channel...

Loading...
Recent Stories

This site, like most other sites on the web, uses cookies. For more information, see our privacy policy. Got it
Close

Email This

This feature is only available to registered users. Register or sign in to use it.