Follow Slashdot blog updates by subscribing to our blog RSS feed


Forgot your password?
The Internet

How the Wayback Machine Works 134

tregoweth writes: "O'Reilly has an interview with Brewster Kahle about how The Internet Archive's Wayback Machine works, with lots of juicy details about how the biggest database ever built works."
This discussion has been archived. No new comments can be posted.

How the Wayback Machine Works

Comments Filter:
  • Google? (Score:4, Interesting)

    by kenneth_martens ( 320269 ) on Wednesday January 23, 2002 @09:25AM (#2887393)
    It's an interesting idea, but the real problem is not storing the 100 TB of data, it's figuring out how to search through it to find what you're looking for. Now, apparently they write a lot of their own software, but it might be better if they could team up with Google and have Google index their sites on a special database. We'd have for regular searches, and for the Wayback Machine's sites.

    Something else I found interesting: according to the article, they "use as much open source software as [they] can." That makes sense when they've got between 300 and 400 computers, and with the number growing all the time. Licensing all those with a non-open OS would be quite expensive.
    • by morzel ( 62033 ) on Wednesday January 23, 2002 @09:53AM (#2887489)
      Please please please please do _NOT_ google it... It was embarassing enough when google acquired dejanews, and put the old usenet archives on-line. :-)
      I just visited some sites from which I hoped that they dissappeared completely from cyberspace. The only defense I've got now are the old cryptic URLs of these monstrosities... Indexing that database would be a disaster, especially with an unusual name like mine...
      (Yes, I was stupid enough to use my real name ;-)
      Damn you, wayback :p

      • Re:Noooooooooo !!! (Score:2, Informative)

        by ichimunki ( 194887 )
        Important to note that they will allow you to "opt out" by using a robots.txt file (not sure what you do if the domain is no longer available).

        Funny part is, they may not have to allow this, except out of courtesy. Apparently libraries such as this can get away with all kinds of stuff that, if done by private individuals with any kind of profit motive, would normally constitute serious copyright violations. (see for information).
      • You know, for sites like the wayback machine, google groups, etc., it would be nice to have another option. Rather than contacting them and saying "please remove this old embarassing crap of mine from your database," it would be great if one could instead tell them "please suppress this stuff for the next 60 years" (and have the request carry forward to whoever inherits the project). That way, people (historians) could still see all my old stupid stuff, but not until I'm dead or too senile to care.
    • Even better, they should use Google to figure out what to cache in the first place.
      The way it works right now, it's more like a mindnumbingly dumb brute-force cache (like I found some of my own irrelevent web pages from years ago).
      While there may be arguments in favour of saving everything and storage may be cheap, while not use Google's ratings to save more relevant/interesting info and ignore the crap...
      • In future years, it may be the crap that people are interested in. For example, researchers looking at old newspapers get a lot of useful information from advertisements, but they wouldn't seem particularly useful at first glance.
    • Yes, searching it is definitely an issue, but I think storing it is also a significant issue. Why did they have to put something important and cool in a proven earthquake area?

      At least the last time we lost the most important collection of knowledge in human history, some human intervention was required--the burning of the library at Alexandria.

      We're collecting a lot of knowledge, but we seem to be forgetting some relevant history. (At least we're avoiding the other California disasters: wildfires, pestilence, mud slides...)
  • by SilentChris ( 452960 ) on Wednesday January 23, 2002 @09:27AM (#2887401) Homepage
    Ok, we have successfully Slashdotted the Wayback Machine. Screw history! :) Let's move on to bigger and better things.
  • Wayback Slashdot [] ...only goes back to 2000? Seems kind of lame when you consider that my little web site goes back to 1998 [].
    • Wayback Slashdot ...only goes back to 2000?

      Wayback [] goes back to 1997...

      • Great, now you're slashdotting pages from the past. Even more ironic, you're slashdotting slashdot itself from 5 years ago. This is why many civilizations ban time travel.
      • You'll notice that Slashdot's logo used to be really big []...
  • Ewwwww! (Score:3, Funny)

    by NeoTron ( 6020 ) <kevin.scarygliders@net> on Wednesday January 23, 2002 @09:34AM (#2887423) Homepage
    And I thought I'd erased all my old embarrassing HTTP handywork....until I discovered my old website nicely archived - bleargh!

    Ah hell, may as well keep it there - it's even got my old web-based Curriculum Vitae on it too - perhaps in some way I've now been "immortalised"?? :)

    I've not touched HTML ever since those first abortive attempts I made 5 years ago, cause I realise now that I'm pretty crap at it - I'll stick to Unix admin, what I know best ;)
  • Wow! You really weren't trying too hard back in the day [], were you Taco?
  • by nurightshu ( 517038 ) <> on Wednesday January 23, 2002 @09:35AM (#2887427) Homepage Journal

    I was glad to see the interviewee was brutally honest about free software -- both its benefits and its drawbacks. Usually discussions among my friends usually degenerate into holy wars, with both of us spouting cliches at one another until we all storm off in huffs.

    Free software can save the world, I think. We just need to realize that it needs a lot more work to get there.

  • by Rentar ( 168939 ) on Wednesday January 23, 2002 @09:43AM (#2887449)
    They don't seem to think the history of their site would be interesting:*/http://web.archive.or g/ [] lredirects you to their index.html! boring!

    Now, that would really be a test for their apps. Same as if Google indexed (entirely).
  • by NOT-2-QUICK ( 114909 ) on Wednesday January 23, 2002 @09:44AM (#2887452) Homepage
    As per the article, Brewster Kahle states that:

    "The idea is to build a library of everything, and the opportunity is to build a great library that offers universal access to all of human knowledge."

    Not only does this sound like a rather far fetched plot from an old StarTrek episode, but it also seems to be an a physical and theoretical impossibility. Even if adequate storage space did exist for such a task (a 10 TB database would be but a small start), I do not foresee any type of technology that could ever adequately capture new data at a sufficient speed to harness that which is human innovation and creativity.

    It is a nice thought, however, and I certainly wish him all the best in her pursuits...
    • Kahle's idea is actually quite reminiscent of Vannevar Bush's seminal 1945 description in The Atlantic Monthly of the memex [], a device that would "give man access to and command over the inherited knowledge of the ages".

      The frequency with which this article (the Bush article, that is) has been cited in hypertext research attests to its importance.
  • Not the biggest DB (Score:5, Informative)

    by costas ( 38724 ) on Wednesday January 23, 2002 @09:44AM (#2887454) Homepage
    100 TBs do not make the biggest DB ever. I am personally working on an 60-70TB ERP system that's also writeable; I am sure there are bigger systems out there (e.g. Wal-Mart's or GM's ERP systems come to mind).

    A read-only DB containing highly-compressible text does not really make for a very challenging datamine. Just because it's on and about the Web and sexier than a stodgy ERP system should not make you overlook the real technology.
    • by limber ( 545551 )
      I am sure there are bigger systems out there (e.g. Wal-Mart's or GM's ERP systems come to mind)

      Just to nitpick, in the interview Mr. Kahle does explicitly mention that the database is in fact bigger than Walmart's. No mention is made of GM's, however.

      "It's larger than Walmart's, American Express', the IRS. It's the largest database ever built. "
      he says. Whether the claim is credible is a different matter.
      • by limber ( 545551 )
        As a side example to this discussion of 'what constitutes a large database', the NOAA's National Climate Data Centre [] maintains a database of digital data of about a petabyte of climatological data. The Centre takes in about a quarter of a terabyte of data *daily*.
      • by costas ( 38724 )
        I find the claim dubious. Bigger than what kind of database? Wal-Mart if famous for tracking every single little thing about their supply chain. Most grocers or hypermart chains do the same. I can easily see, say A&P or Tesco or Carrefour having multi-TB DBs, even petabyte DBs.

        Also, the size is not the only thing that defines a database installation: numbers of simultaneuous users or concurrent transactions, read or write access, ability to rollback, quality of service standards are way more important in my book (and also for most big companies). Part of the reason DBs in that size range are rare is exactly that current technology does not scale up to those levels while maintaining rollbacks, read-write and fast user response.

        I like the Wayback machine, but to compare it to a proper database is ludicrous. EMC or Veritas will give you much more for their 100TBs of storage than 400 x85 PCs... instant backups for one and way larger MTBF.
    • I'm not sure how much storage Deep Thought had, but certainly the computer it was priviliged to design has more than any database on this planet - by definition, since that database is simply a subsystem! Even more impressively it compresses down to just two (base 13) nibbles: 42.
    • You're right, the Wayback machine is not the largest collection of data -- not even the largest collection online. I work with the USGS's catalog of satellite data. They have over 300 terabytes of satellite imagery, and the collection is growing at a rate of about 1 terabyte per day.

      The USGS collection comprises multiple instruments, but Landsat 7 is a big one, contributing about 100 terabytes that's searchable online. []

      Perhaps 'Largest TEXT Database' would be a better description of the Wayback Machine?
      • Wow. How do you even deliver 1TB/day to a database? That's more than enough to saturate a 100Mbps network, 24 hours/day. And how do you index it, so it can be usefully retrieved? And how do you store it?

        That is a truly astounding amount of data. It's like receiving Google's entire database every week, several times over.
  • by shic ( 309152 )
    So... would I be right in thinking that the "Wayback Machine" is an (admittedly large scale) exercise in network database programming (of the style popular pre Codd and his relational model?) I am tempted to question if this is indeed the biggest database ever built - I suppose it depends upon definitions - but to my mind a database should be general purpose... whereas it appears to me that this project is basically a large-scale single index.

    I also wonder if it would be appropriate to call this the largest project of it's kind - for example - while Google stores less data, I suspect it supports a higher query rate... how exactly do you intend to measure scale... if it is in terms of computing power is it relevant that Google already have thousands of Linux server nodes?

    That said - I think it is an exciting project in its own right. I hope and expect this offering to become a significant information resource in years to come.
  • What? (Score:2, Funny)

    by Ezubaric ( 464724 )
    I thought it was going to tell us how Mr. Peabody's wayback machine worked. You know, like the flux capacitor diagrams that made everything clear . . .
  • by CDWert ( 450988 ) on Wednesday January 23, 2002 @09:53AM (#2887487) Homepage
    Id say is pretty amazing, I actually was able to retreive content I thought lost years ago.

    My sites go back to 95, and yep theyre archived starting 96, this is too cool.

    I wonder how much of the goverments docs that were pulled off post Sept 11 are still on this ?

    A really funny note is it seems like all the p0rn is intact staring in 96, gotta archive the porn.

    But seriously , I was unaware of this, Im gonna use this thing like hell as a sales tool if nothing else. Its also great to find certain content thats been pulled.
  • sigh (Score:2, Insightful)

    by __aahlyu4518 ( 74832 )
    This world needs more people like that... driven to make this world a better place... and having fun doing it, being proud of what they're trying to accomplish... This interview sent shivers down my spine... These are the kind of interviews that inspire people... It makes me think about humanity a little less sceptic... There's still hope.
  • The article was good with all it's warts and gems. I take it more as a testament to the human spirit. I seriously doubt it's the largest database, though it might be the largest publicly accessible database. I'm sure the NSA could easily dwarf their database considering how much data they collect from around the world every day.
  • by wayn3 ( 147985 ) on Wednesday January 23, 2002 @10:01AM (#2887512)
    You buy from EMC a terabyte for maybe $300,000. That's just the storage for 1 TB. We can buy 100 TBs with 250 CPUs to work on it, all on a high-speed switch with redundancy built in.
    Interesting quote. Mr. Kahle addresses something I've been wondering for a while -- are storage area networks really worth it? Or is he ignoring the costs of maintenance and manpower to keep these things afloat?
  • by Karma Star ( 549944 ) on Wednesday January 23, 2002 @10:02AM (#2887514) Journal
    Seeing that they cache webpages from other sites, I wonder how long it will take before another company sues them?

    Also, I wonder what their criteria will be for "submissions"? 1 month? 1 year?
    • Google, and others, also cache a lot of content. If a web provider doesn't want their stuff cached on The WayBack, all they have to do is include the no bots code in their html.
    • Child! Child! They do not sue you right away -- and they can't. First they send you a cease-and-desist order and you evaluate their claim.

      But Brewster answers your question in the interview himself on the second page:

      Koman: What about the question of rights? I just wrote about Lawrence Lessig's book on intellectual property. Surely the publishers and the television networks and the record companies aren't willing to let you keep a copy of all of their stuff?

      Kahle: All we collect for the Web archive are sites that are publicly accessible for free, and if there's any indication from the site owner that they don't want it in the archive, we take it out. If there's a robot exclusion, it's removed from the Wayback Machine. Over the years, people would notice these things in their logs and would say, what are you doing? And we'd explain what we're doing -- building this archive and donating a copy to the Library of Congress, etc., etc., and 90% of the time they say, "Oh, that's cool, you're crazy, but go ahead." About 10% of the time, they'd say, "I don't want any part of it," and we instruct them on how to use a robot exclusion and they're taken out of history. That seems to work for everybody at this point. People are really excited about this future that we're building together.

  • by Tazzy531 ( 456079 ) on Wednesday January 23, 2002 @10:02AM (#2887516) Homepage
    The interview talked a little about throwing more machines on when the demand deems necessary. I wonder if it is possible to do this over the internet? I mean, I'm seeing something along the lines of SETI, where millions of people worldwide donate their unused processor power. Would it be possible to distribute the searches to remote computers over the internet in real time?
  • by Tazzy531 ( 456079 ) on Wednesday January 23, 2002 @10:14AM (#2887571) Homepage
    A number of you have asked whether the websites taken down since 9/11 are available on The answer is yes. One example is:

    DC Air National Guard on Archive []

    Same Page - 404 []

    One of the conspiracy websites that I have read was saying that combat airplanes, normally on 24 hour alert, at this base should have and could have prevented the plane from entering the restricted airspace in DC. They were saying that this site was removed because it provided evidence that somebody dropped the ball.
    • Living a five minute walk from the Pentagon I sort of followed the "WTF WHERE WERE THE JETS" stories. I was wondering myself. Turns out the Jets were launched and sent towards New York as air controllers were searching for additional planes. Then they found one...heading for Washington. The Jets arrived within range about 5-10 minutes after the first Jet crashed into the Pentagon.
      • I am in no way saying that what you heard might not be true, but read this analysis [] of it. I don't totally agree with their analysis, but it sure does make you think. According to them, no jets were scrambled until after the fact.

        Make your own decision...
      • It was the ol' duck and roll, the ol' feint and lunge, the ol' slice and jab. I don't know if they planned it that way, but that's how it worked. It was a strategy error that a typical Risk player would make. They should have kept two planes in the launch area, and sent just two to NYC. They forgot to cover home base. It's like they sent the catcher to third base, or brought the safety on a blitz.
  • " Question:
    Are you violating copyright laws?
    About the Internet Archive

    No. Like your local library's collections, our collections consist of publicly available documents. Furthermore, our Web collection (the Wayback Machine) includes only pages that were available at no cost and without passwords or special privileges. And if they wish, the authors of Internet documents can remove their documents from the Wayback Machine at []."

    I don't really think that they're neccesarily right about this. I'm glad they've got the archive up, and I think it's dandy, but it seems like the copying and reposting of other's materials is a suspect practice. This will end up in court as soon as something that someone removed from their own webspace re-appears historically accurate here. I'd guess some liable suits will be the first...

    • I agree. This does walk the gray area of legality.

      For example, I took the Term of Use Agreement [] from
      (B) CNN Interactive contains copyrighted material, trademarks and other proprietary information, including, but not limited to, text, software, photos, video, graphics, music and sound, and the entire contents of CNN Interactive are copyrighted as a collective work under the United States copyright laws. CNN owns a copyright in the selection, coordination, arrangement and enhancement of such content, as well as in the content original to it.
      Subscriber may not modify, publish, transmit, participate in the transfer or sale, create derivative works, or in any way exploit, any of the content, in whole or in part. Subscriber may download copyrighted material for Subscriber's personal use only. Except as otherwise expressly permitted under copyright law, no copying, redistribution, retransmission, publication or commercial exploitation of downloaded material will be permitted without the express permission of CNN and the copyright owner. In the event of any permitted copying, redistribution or publication of copyrighted material, no changes in or deletion of author attribution, trademark legend or copyright notice shall be made. Subscriber acknowledges that it does not acquire any ownership rights by downloading copyrighted material.

      I'm sure CNN isn't the only site that has this type of Policy.

      If they ever do get sued and are held liable for these copyrighted materials, it would be at a major lost to the global community. The internet is a part of our history and has a history of its own.
      • They can stick all the policies they want on it, but I doubt that they have any legal force. Normal copyright laws and exculsions would apply, and, since there's no agreement or license, regardless of what they stick in the small print, nothing extra applies. In fact, as I read more closely, it's right there - "Except as otherwise expressly permitted under copyright law".

        To my knowledge (limited), publicly available archives have always fallen under fair use.

        Thats not to say that someone won't sue, but, considering that they seem to be more than willing to pull things at owner request, I doubt they'll ever end up in court. And if they do, they'll probably win.

      • That's a policy in English for human beings to interpret. If they want to stop browsing automatons from cacheing/indexing, the Robots Exclusion Standard is the common and almost always adhered-to standard (except for spambots, of course).
    • In order to resolve the copyright issues, and in order to preserve the public domain on the Internet, the Internet Archive has filed a friend of the court brief in our court case against the Sonny Bono Copyright Term Extension Act. See []

      As of today, the Supreme Court hasn't decided what to do with the appeal. Stay tuned to openlaw.
    • I'm pretty sure they're not far off. I spent some time this morning reading documents at and I got the impression that under most circumstances, libraries could do all kinds of things that regular people couldn't.

      I also note that is owned by Alexa Internet, and according to " © 1996-2001, Alexa Internet, Inc. Service provided by Alexa Internet. A wholly owned subsidiary of " So what we really need to be concerned with is that this whole thing will be patented and we'll have to get a license from Amazon to have Linux clusters or online databases.
    • Maybe technically, but as long as they are willing to quickly pull a site from the archive, and work with the site owner to prevent future content from being archived, I can't see how it will ever end up in court.

      Lawyers are expensive. Nobody is going to file suit without first sending a "please remove my data from your archive" letter.

      And if some site owner is completely stupid and sues first, they can simply pull the pages from the archive, then show up in court and say "we deleted their pages as soon as we were made aware of the problem", and the judge will dismiss the entire case.

  • by rkgmd ( 538603 ) on Wednesday January 23, 2002 @10:16AM (#2887594) l Apparently, walmart's is 24TB, and the entire www index as of 1999 was only 6TB.
  • Hardly the biggest. (Score:2, Informative)

    by dmd ( 404 )
    The biggest database ever? 100TB? Hardly.

    I worked at a large pharmaceutical company for two years (known internally as the Squid), and supported a 380TB protein interaction database (Oracle) and a 260TB SAP-backend database (Informix + custom).

    Certainly Wayback's database is large, and certainly it holds far more varied information and appeals to a far larger audience, but by no means is it the biggest. I'm sure there are databases that made the ones I worked on look puny by comparison.
  • Even though this has been available since Oct, it's the first I've seen of it. I think it's a great resource. Long dead sites that are no longer there now can be found for historical purposes. The interesting thing is that the links on the page are also updated to link to the archived versions. What I found it useful for was building a history page of what my site looked like over the years. Lots of great uses for this so hope it stays up!

  • So the article says in one place that they wrote their own operating system, and in another that they use linux (or BSD, I forget which).

    So which is it?
    • Re:Operating system (Score:2, Informative)

      by madfgurtbn ( 321041 )
      They use different OS's for different purposes within the system. The so-called OS they wrote is described in the article. It's a collection of tools for controlling their parallel computer, which is a collection of many inexpensive computers running the BSD and Linux OS's you talk about.

      The interviewer is the one who describes it as an OS. The interviewee expains that the real breakthrough is that with their tools an ordinary programmer can operate in a parallel computing environment-- you don't need a specialist in parallel computing anymore. Which leads to the conclusion that relatively small institutions on relatively small budgets can build enormously powerful computers with massive storage.
  • DBMS and model? (Score:3, Interesting)

    by leandrod ( 17766 ) <l@dutr a s .org> on Wednesday January 23, 2002 @10:44AM (#2887740) Homepage Journal
    But what is the DBMS? Is the database relational? How it was modelled?
  • by Proud Geek ( 260376 ) on Wednesday January 23, 2002 @10:50AM (#2887764) Homepage Journal
    I once worked on a site with a 25 year old database that was much larger.

    The ancient magnetic storage took up several warehouses. Beat that, for biggest database ever!
  • by cheese_wallet ( 88279 ) on Wednesday January 23, 2002 @10:57AM (#2887799) Journal
    Pretty decent read, but one thing they said got me thinking a little bit.

    They said that at Thinking Machines they built a super fast computer, but it required a new way of thinking about things in order to program it. And then they called this a mistake, because they couldn't attract any customers.

    This seems like a real problem that would lead to technological stagnation. At least from a market place point of view.

    It is kind of similar to a company making games off of pre-existing engines, like quake, instead of some new non-quake compatible engine.

    Or everybody making x86 compatible CPUs.

    It also seems that when a company does come up with some new way of doing things, they get burned, and it is the second generation of companies that pick up the torch that make the money. So nobody wants to be that first company, they are all waiting for someone else to break the ground.

    Maybe the only people/companies that come up with new stuff are the ones that are insanely rich, and won't get hurt by doing something new, or the insanely poor who have nothing to lose anyway.

    I can't help thinking that this clustering boom going on is just like what 3dfx was trying to do. The difference right now is that clustering actually *does* outperform the super fast single chip. I wonder when technological advances will change this fact.
    • They said that at Thinking Machines they built a super fast computer, but it required a new way of thinking about things in order to program it.

      That caught my eye too, for a couple of reasons. First, I think in a way, Thinking Machines was a victim of its own success. At first, people thought they were nuts building computers with 65,536 processors in them. Then, when people realized hey, that's actually pretty handy, suddenly there was pretty intense competition for a relatively small market.

      Second, what was a "new way of thinking" in the late 80's and 90's when Thinking Machines was doing their stuff, may now have become the norm (or at least "a norm"). A lot more people now think about parallel processing. So you're right, the first guy to do something may very well get burned, and after burning, can watch everyone else succeed.

      Third, I think Brewster's sound bite about TMC (Thinking Machines Corp) came off sounding like he was selling them short. I used TMC's Connection Machines, along with parallel machines from other companies (one called Masspar comes to mind). The TMC software was pretty slick in its day. One really great feature it had, which others lacked, was that in the beginning of your program, you specified a "virtual processor geometry", i.e. you told the system "I want to run this program on an array of 1000x2000 processors". The software would then figure out, oh, this machine only has 65536 processors, not 2 million, and it would map the 2 million virtual processors onto the 65536 physical processors, and the programmer didn't have to worry about doing that mapping. It sounds simple, but again, I used other big parallel machines at the time which couldn't do it, which made things a pain (writing a chunk of code in every programming to do that mapping myself). You could take your code which requested 2 million virtual processors, and run it on Connection Machines with different numbers of processors, and it would Just Work.
  • This site [] has archived television from all over the world on September 11th-September 18th.

    I've been pretty jaded and unpatriotic/anti-war about the whole thing, but I can't help but admit I still get creeped out watching the footage.
  • by dood ( 11062 ) on Wednesday January 23, 2002 @11:20AM (#2887915)
    Slashdot looks the exact same it did 5 years ago!

    WHEN is this site going to be updated? Forget the wayback machine, if I want ancient web history I visit slashdot.

    • Rob (CmdrTaco) has always said that if you come up with a better design, email him and he'll put it up if he thinks it's worthy. So instead of complaining, go redesign Slashdot yourself!

      It would sure be a cool thing to be able to say you designed Slashdot. :)
  • by grub ( 11606 ) <> on Wednesday January 23, 2002 @11:32AM (#2887994) Homepage Journal

    From the article:
    How the archive works is just with stacks and stacks of computers runnning Solaris on x86, FreeBSD, and Linux, all of which have serious flaws, so we need to use different operating systems for different functions.

    The man puts bias aside and uses various OSs in areas in which each performs well. A real, tangible project like this is worth more than any amount of drooling zealotry.
  • Arg. Where's Google's cache? Oh.. wait.. nevermind.
  • by selan ( 234261 ) on Wednesday January 23, 2002 @11:37AM (#2888027) Journal
    Having so few transactions for a database of this size probably helps them run without needing large expensive machines. Many VLDBs [] support thousands of transactions per second. I found a list here [] of top ten winners of a very large database scalability contest. The winner for peak performance was something like 20,000+ TPS.
  • imagine what beowulf joke we can put...
  • This brings to mind a serious question. What data, if any, do we throw away? With ever expanding storage capacities it's getting easier and easier to just keep stuff, than to sit down and figure out what you want to throw away.

    In the past media degeneration and obscelecense over time have made the decisions for us. But going forward we will have massive distributed, redundant data stores, with geographically remote backups. The data isn't going to go away unless we tell it to.

    Freenet addresses this problem by culling the less popular data (not actively, but as a end result of its caching policies) - but this has the unfortunate effect that important data can get lost. Not a desirable behavior for corporate data.

    • : In the past media degeneration and obscelecense
      : over time have made the decisions for us. But
      : going forward we will have massive distributed,
      : redundant data stores, with geographically
      : remote backups. The data isn't going to go away
      : unless we tell it to.

      GOOD! And there is something *wrong* with this??
      Seriously -- the Internet should have been designed like this from the start. Don't ever throw away, simply classify and organize. In a throw away world, we already destroy too much to afford loosing what's on the Internet.

      Given, there is lots of junk and what some might consider noise, but... (as the saying goes) some peoples junk is others treasure.
  • With an appropriate robots.txt [] file, a site's listing can be stopped from showing up [] on the Wayback archive. Interesting.
  • I was kinda bummed to see that and yielded:

    Blocked Site Error.

    Per the request of the site owner, is no longer available in the Wayback Machine. Try another request or click here to see if the page is available, live, on the Web.
  • I hope people help them out, they have already brought me back to some cool stuff.

    This is a noble cause.

  • That must be like 99.99 TB of pr0n!!!
    • Well regardless, I love the fact that they keep binaries such as hard to find .tgz's and shelved projects. It's amazing how some things just seem to fall off the net, and then are almost impossible to locate.

      I so hope, that they don't have to contend with any legal troubles which might interfere with the initiative. If so, they should move the archive to an island, move the servers to switzerland, or a boat somewhere at sea, where they can have legal immunity. This is too great a cause to have tangled up in a brainless legal mess.
  • I can't seem to get the slashdot pages archived to load.
  • by joshuaos ( 243047 ) <{moc.tnemodeerf} {ta} {soroboruo}> on Wednesday January 23, 2002 @02:07PM (#2889074) Journal
    I spent the summer on the road, and when I settled down the for the cold months, I was quite sad to see the the Misanthropic Bitch [] appears to have vanished. This made me very sad. Today, when I read this article, I was delighted to find that all of dear bitch's articles are archived.

    I think this is a fabulous project, and I hope it does well. However, I think that the notion of such a centralized database will begin to become unrealistic. I think peer to peer projects are the future, and I can see a day far in the future when the database layer comes down and inhabits the filesystem layer and all the databases on the internet can talk to eachother, and in a sense, the net becomes a giant database that anyone can contribute to.

    Cheers, Joshua

  • by for(;;); ( 21766 ) on Wednesday January 23, 2002 @03:00PM (#2889477)
    Hot damn! Their movie archive has a downloadable version of the short they showed on MST3K prior to "'Manos:' The Hands of Fate."

    "Ma'am, did you realize that Chevrolet has an important plan for your life?"

  • by russ-smith ( 126998 ) on Wednesday January 23, 2002 @03:11PM (#2889554) Homepage
    The majority of information being collected by is covered by copyright law. It is up to to get permission before they republish the information. If you look at the Archive web site they run banner ads for the Alexa toolbar. This Alexa service provides the marketing with information somewhat similar to the Nielson ratings for TV. has received complaints about their service contrary to the statements made in the published article. has refused to respond to any meaningful way to these issues. is trying to put burden on the publisher to determine that The Archive is publishing it, find it within TheArchive web site and then provide them a notarized statement. see their FAQs at ml?index=2 [] and ml?index=26 []

    The claims made in these faqs are just not consistent with the law. Are they going to repost everything that was available on Napster?

    They also have some problems with their algorithm so that some domains that are redirected fool their algorithm into associating content with a site that was never actually associated with the site. To try to find copywritten works would be a nightmare. has refused to respond to any of these issues and, in fact, are lying about it if the quotes in the article are factual.

    Russ Smith

  • slightly off topic but check out ID Webpage circa Dec 1997 and then look at it today. The work that must have gone into update that baby (;
  • The US federal gov't has larger databases than this. They just don't talk about them too much.

    BTW, looks like the Wayback Machine has been slashdotted.
  • I wonder if Google indexes this site. Think of the ramifications on finding the revellent inforation. Paradoxs galore!!
  • wow!
    just imagine a beow.......

    ah. nevermind...
  • well, look at that... it has tons and tons of lost and forgotten porn... time to buy that new 160 gigabyte hard drive!
  • Has anyone tried this site in Konqueror? There is a floating link to the home page on top of the results that says search in progress, but never goes away?
  • index it, split the database horizontally each split on a different machine.
    I would have to look at the database to tell you how big each split would ned to be, and how much power the machine housing each split needed to be, as well as other details.
    depending on certian variables, I might even consider splitting it accross a cluster od some sort. but its hard to say without a real look at the structure.

Loose bits sink chips.