Wayback Machine Way Out Of Lawsuit

from the going-wayback dept

Two years ago, we wrote about how a law firm was upset that its counterpart in a trademark case had used the Internet Archive's Wayback Machine to view some old websites for the case. Apparently the law firm had set up a robots.txt file to block the Archive and other sites from searching and archiving those pages. Last summer, the Internet Archive was sued over the issue. Now, comes the news that they're no longer a part of the lawsuit, though the original law firm is still accused of "hacking" and violating copyrights to view the old websites. Unfortunately, there are no details on how the settlement was reached or what it entailed. Internet Archive founder Brewster Kahle says that the accessibility of the pages was a "temporary glitch" that has since been fixed. That's good, but unfortunately leaves a lot of the other issues related to this case unsettled. Hopefully some of those will come out of the final result of the other part of the case.
Hide this

Thank you for reading this Techdirt post. With so many things competing for everyone’s attention these days, we really appreciate you giving us your time. We work hard every day to put quality content out there for our community.

Techdirt is one of the few remaining truly independent media outlets. We do not have a giant corporation behind us, and we rely heavily on our community to support us, in an age when advertisers are increasingly uninterested in sponsoring small, independent sites — especially a site like ours that is unwilling to pull punches in its reporting and analysis.

While other websites have resorted to paywalls, registration requirements, and increasingly annoying/intrusive advertising, we have always kept Techdirt open and available to anyone. But in order to continue doing so, we need your support. We offer a variety of ways for our readers to support us, from direct donations to special subscriptions and cool merchandise — and every little bit helps. Thank you.

–The Techdirt Team


Reader Comments

Subscribe: RSS

View by: Time | Thread


  1. identicon
    Anon, 25 Aug 2006 @ 12:38am

    "temporary glitch" doesn't really explain what happened. sounds like they screwed up and maybe paid out to make the case go away.

    link to this | view in thread ]

  2. identicon
    Anonymous, 25 Aug 2006 @ 5:15am

    Something must be missing here

    Either something is missing here, or I misread the article.

    How could the Wayback Machine make web pages available that it shouldn't have indexed in the first place, unless it completely ignored the robots exclusion file on the original site? That's not a "temporary glitch".

    The Wayback Machine wiki states "The archive also maintains the "Wayback Machine", with content from Alexa Internet." Alexa is a commercial data mining operation. I suspect some site managers might not want Alexa indexing their site if they knew this.

    link to this | view in thread ]

  3. identicon
    Sanguine Dream, 25 Aug 2006 @ 5:32am

    So lets get this straight...

    Two firms are going at in a lawsuit.

    Firm A uses the Wayback Archive to collect evidence.

    Firm A then sets up a bot exclusion file to stop others from getting to the old sites they were using for evidence?

    Help me out here.

    link to this | view in thread ]

  4. identicon
    Azuravian, 25 Aug 2006 @ 6:48am

    Re: So lets get this straight...

    I think it's more like this.

    Firm A collects information from old websites.

    Old websites are owned by firm B.

    Firm B files lawsuit including Wayback because old websites have robots.txt setup to exclude them from inclusion in Wayback archives.

    link to this | view in thread ]

  5. identicon
    Anonymous Coward, 25 Aug 2006 @ 9:19am

    I still don't get it...

    So explain to me how this has any legal precedent. The way I interpret it is as such:

    I create a document that I distribute freely to anyone that asks for it, with a small note at the bottom that says, "If you are a robot, don't store the contents of this page."

    I then have a second thought about this document, and I destroy the original.

    Somebody finds a copy of this document, in this case from a robot (although I don't see any relevance in how they obtained the copy, except for my small “do not store this page” note), and I get mad and sue them?

    The document(website) was public at one time. Anyone could have made a copy at any point. Just because a robot made the copy it is a legal issue? What if I had a cached copy, or I printed the website before it was taken down? Could they sue me?

    This could be a whole new way for me to make money. Forget my 9 to 5, I just need to track down all the copies of my old websites and sue the people that made a printout or copy!

    link to this | view in thread ]

  6. identicon
    Joe Smith, 25 Aug 2006 @ 12:15pm

    Like Virginity

    confidentiality can only be lost once.

    Seems to me that the Plaintiff was trying to allege that certain information was confidential and the defence was trying to show that the information had previously been made available to the world through the web.

    The question I have is: What information could there possibily be on that old web page which the Plaintiff could properly seek to suppress?

    link to this | view in thread ]


Follow Techdirt
Essential Reading
Techdirt Deals
Report this ad  |  Hide Techdirt ads
Techdirt Insider Discord

The latest chatter on the Techdirt Insider Discord channel...

Loading...
Recent Stories

This site, like most other sites on the web, uses cookies. For more information, see our privacy policy. Got it
Close

Email This

This feature is only available to registered users. Register or sign in to use it.