Using Google to Calculate Web Decay 209
scottennis writes: "Google has yet another application: measuring the rate of decay of information on the web.
By plotting the number of results at 3,6, and 12 months for a series of phrases, this study claims to have uncovered a corresponding 60-70-80 percent decay rate.
Essentially, 60% of the web changes every 3 months." You may be amused by some of the phrases he notes as exceptional, too.
At last! (Score:1, Interesting)
Google's collection of the data (Score:2, Interesting)
Surely this can't be true. Check Google's cached pages - see the dates on there?
Google is turning into another history book [archive.org].
History book? Not as far as I can tell . . . (Score:3, Interesting)
Googlebot Visits Monthly (Score:2)
Are google claiming that they can check through the entire internet inside a timescale of 3 months, ready to check through again at the start of the next quarter?
I don't know if that's all that far-fetched. I know Googlebot last hit my site on April 7th, crawled every page in my domain over the course of 12 hours, and current searches of their cache show content I'd updated at that time. They seem to visit every month or so.
Perhaps it's based on the traffic they detect to a given site through their CGI redirects... but I'm not a large site, my primary webserver is a Pentium 90. :)
crawl4.googlebot.com - - [07/Apr/2002:13:36:32 -0400] "GET /broken_microsoft_products/ HTTP/1.0" 200 128854 "-" "Googlebot/2.1 (+http://www.googlebot.com/bot.html)"
Re:Googlebot Visits Monthly (Score:1)
crawl4.googlebot.com - - [07/Apr/2002:13:36:32 -0400] "GET
wow, not only are you running your domain off a pentium 90, but you also have reverse DNS lookup turned on in the logs... that's gotta be giving you a decent preformance hit, no?
Re:Googlebot Visits Monthly (Score:2)
wow, not only are you running your domain off a pentium 90, but you also have reverse DNS lookup turned on in the logs... that's gotta be giving you a decent preformance hit, no?
Well, it doesn't actually handle DNS; that's felix, an old 486DX-33 running FreeBSD, port-forwarded behind my gateway (I've only got the one IP address). But yeah, I'm sure each logger thread gets held up waiting for resolution.
More impressively, dynamic content. (Most of the pages are generated dynamically as shtml through the x-bit hack; nothing sophisticated, mostly just inserting templates and stuff for color scheme because I'm too lazy to type long BODY tags) And anywhere from 2,000 to 5,000 hits per day. And only 48 megs of RAM. And it's a popular Linux distro's default kernel, not recompiled for that machine. Even so, it hardly ever breaks a sweat.
As you can tell, it's like, zero performance tuning. But it still cranks out a SETI@Home unit every day or two.
As for reverse DNS itself, yeah, I like it. :) It's a nice luxury.
Not exactly decay... (Score:4, Interesting)
That said, this is one of the many ways in which Google is an invaluable tool for research. Not just finding information, but generating it. Thanks Google!
Re:Not exactly decay... (Score:1, Insightful)
It's the medium size businesses that don't seem to be grasping the web and the fact that you need to have a site that is dynamic in so far as it keeps people interested and possibly entertained.
I'm lucky in that the company I work for is a small firm and a publisher so we have daily news content and well as on-line versions of our weekly and monthly publications (HTML and PDF downloads!) being uploaded all the time - so our web traffic is growing constantly - slowly but it hadn't seen a decline in the past two years.
M@t
bill gates sucks... (Score:2, Funny)
Re:bill gates sucks... (Score:1)
I say it should be admitted as evidence in the recent court hearings as proof of what the public thinks of Mr. Bill (and his company which is an extension of his being). =)
Re:bill gates sucks... (Score:5, Interesting)
Let's face facts here. We might hate Microsoft, but the vast majority of people do not. Good? Bad? Indifferent?
Kierthos
Re:bill gates sucks... (Score:1)
Kierthos
Re:bill gates sucks... (Score:4, Insightful)
It's an indicator of the dubious kind of context in which one finds such rash statements.
Re:bill gates sucks... (Score:1)
Kierthos
even Bill thinks he sucks more than ever (Score:3, Interesting)
Google SOAP thing for compare-stuff is in the pipeline...
An ever-eveolving creature? (Score:1)
Information vs WWW (Score:1, Interesting)
In fact, Information Superhighway would be a great data point for this subject. Another consideration, which would be difficult to accomodate, is the reality of mirrors and shuffling pages to different URLs.
Most importantly, I strongly hope that your "interesting application" never gets implemented, because I can see no application of the resulting data that doesn't make my blood run cold. Psychological Warfare and hostile advertising are the bane of the Post-WWII US, and (likely) the world. Propeganda is a pernicious technology, and I fear further development in this area.
Okay, I'll admit that was a touch trollish. Because the Psych. Warfare genie was already released from it's NAZI bottle and invited into the US (along with other valuable sciences), it's a little late to advocate repression of this technology. Yet I still reel from my country's increasingly malevolent commercialism aspects, which have spun off from Capitalism without any of Capitalism's redeeming social aspects. I almost want to become a socialist, until I consider that this state of affairs sprung from the National Socialist state.
In any case, while the WWW may be evolving, is certainly isn't in the Darwinian sense that was likely intended. Vestigal Geocities homepages long abandoned are plentiful, and are less temporary, giving search engines a better shot at crawling than dynamic, or "living" news portals. This sickly "creature" is more of a construction than the product of evolution (unless you consider pre-Charles Darwin senses of the word). If you want to research the nature of information and survivability/mutability, the Freenet Project would provide a much more fruitful environment, if it ever reached widespread useage. I would have less strenuous objections to classifying the Freenet an "ever-evolving creature".
blessed (Score:4, Funny)
I don't look forward to that day.
Long live cheese and cheese makers!
Blessed are the cheese makers (Score:2)
It's not meant to be taken literally. It refers to any manufacturers of dairy products.
-- Monty Python, Life of Brian
Web Death (Score:4, Interesting)
Re:Web Death (Score:1)
Can't find any links unfortunately (the results of search for anything involving the words "half-life" tend to be somewhat skewed...)
Re:Web Death (Score:4, Interesting)
Also this, which is just a link from the previous article. [unl.edu]
Easy! :)
(web's half-life -game -unreal -counter -gamers)
Re:Web Death (Score:2, Funny)
heh (Score:2)
Re:Web Death (Score:2)
now its all about finding open relays to megaphone your get rich quick idea that you copied from some other guy to 30 million people, praying that you get at least 40 back. course if you decide to bite just to mess with them, you find that they dont even check the box. whats the point? arggvhhh its just frustrating. its completely trashed the fun of having email. and the web.
and the ghost of the old web, the one with low noise, is not viewed as dead, merely its soul is an HTTP redirect to someone's digital billboard, completely unrelated and unwanted.
Re:Web Death (Score:2)
I suppose the rate at which new links are created is roughly a positive coefficient that outweighs the negative coefficient associated with death of a link.
Reminds me of calculations for population growth with k_growth and k_death.
So, two questions:
baseline (Score:1)
Re:baseline (Score:1)
Certainly not as quickly as it would have without your reference, God willing.
Re:baseline (Score:1, Offtopic)
A few things... (Score:1)
First of all, he showed very little of his actual data. This makes it difficult to tell if his interpretation is correct.
Thirdly, what the heck was this guy smoking when he came up with search phrases. Most of these phrases seem to be tangental to the main purpose of most web sitees on the internet.
Finally, Timothy, why didn't you put the foot icon by the story?
Re:A few things... (Score:1)
Justin
One of the flaws (Score:2, Troll)
But, it is interesting to see his results. I can only imagine that if Archive.org [archive.org] did a study like this, they would be able to make a more legitimate conclusion. Perhaps some collaboration is in order?
Re:One of the flaws (Score:1)
-Berj
Another flaw (Score:1)
Obligatory Full Text (Score:5, Informative)
Web Decay
by Scott Ennis
4/26/2002
Knowing how anxious most companies are to keep their web content "fresh," I was curious how "fresh" the web itself was.
In order to come up with a freshness rating for the web you need to sample a very large number of pages. Not wanting to do this, I opted to use the Google search engine as a method for reviewing the web as a whole.
My hypothesis is this: By searching Google using some common english phrases and returning results at various time points, a baseline can be reached for the common rate of freshness of overall web content.
I took the total number of pages found for each given phrase at 3, 6, and 12 months. I calculated a percentage for each of these points based on the total number of results found with no date specified.
For example: Phrase 3 mos. 6 mos. 12 mos. Total
buy low sell high 4700 5470 6200 7830
60% 70% 79% 100%
Note:
This method excludes any pages which are not text and more specifically, not English text.
This method relies on a random sampling of phrases.
Using this methodology I determined that the average rate of decay of the web follows a 60-70-80 percent decline at 3, 6, and 12 months.
Therefore, If a company wants to maintain a freshness rate on par with the web as a whole, their site content should be updated at the inverse rate. In other words:
60% of the site should change every 3 months
70% of the site should change every 6 months
80% of the site should change every 12 months
The only way to do this effectively is to either have a very small site, or have a site with dynamically generated information.
The following graph shows the decay rate for a few phrases. I selected these phrase to display because of their unique characteristics.
bill gates sucks--This phrase had the lowest decay rate of any phrases I searched.
life's short play hard--This phrase had the greatest decay rate of any I searched (note: this search was also very small).
blessed are the cheesemakers--This phrase was relatively small, but demonstrates that quantity of pages may not be important in determining decay rate.
late at night--This phrase returned the highest number of results of any I searched and yet it also adheres closely to the 60-70-80 rule.
Conclusion:
Web content decays at a uniform, determinable rate. Sites wanting to optimize their content freshness need to maintain a rate of freshness that corresponds to the rate of web decay.
Mirror Site (Score:2, Informative)
Free hosting is a bad bargain (Score:4, Insightful)
Not to mention that well over 50% of the time any search engine result that points to Angelfire in particular points to a 404 Not Found. This is much more than what I experience with other sites. Do their users get kicked off often, or just go away, or what? I don't even bother clicking on those results unless it looks like the content is truly compelling. And thank God for Google's cache.
I can understand if some truly can't afford hosting, but even for these people, even Geocities is much better!
Somehow I doubt the majority of those people using Angelfire, Tripod, etc can't afford hosting.
Well, after the dot-com world gets a little more squeezed, those sites may no longer exist. Too bad that many people won't bother rehosting their content and will just drop off the web.
olm.net [olm.net] offers Linux [linux.com] based hosting for under $9/month. No I don't work for them, but I am a (satisfied) customer.
$9 a month - and you won't piss off your users.
(Yes I know their other packages are more - but the $9 a month package is better than any of the free services)
Don't EVEN get me started on organizations and commercial BUSINESSES (ack!) that use free hosting - that is so unprofessional. I don't think I'd want to do business with a company (even a local store) that wouldn't/couldn't pay $9 a month to have a less annoying and more reliable website.
Of course, some of the content out on the Web isn't even worth $9/month, heck some of it has NEGATIVE worth.
Re:Free hosting is a bad bargain (Score:2)
Study: World Wide Web sites and page persistence (Score:5, Interesting)
Digital libraries and World Wide Web sites and page persistence [informationr.net]
Sig: What Happened To The Censorware Project (censorware.org) [sethf.com]
Intermittence (Score:4, Funny)
For example, most web pages linked to in slashdot articles.
Credibility? (Score:2, Interesting)
From the evidence, he searched for very few phrases. The sample size is way too low to be representive of the web - which some estimates put at several billion more pages than there are people on the planet! There are no signs of more than about 5 different phrases being searched for here..
Can a few simple searches on Google really generate a large enough sample to draw such large conclusions?
The report is one page long, hosted on Angelfire. There is no substantial data to back up his claims. Is this report reliable in any way?
I'm amazed this got posted on the front page of Slashdot..
Re:Credibility? (Score:1)
Re:Credibility? (Score:1)
Kierthos
Re:Credibility? (Score:2)
I calculated a percentage for each of these points based on the total number of results found with no date specified.
IMHO, This is a bit vague to be called anything but conjecture.
Re: (Score:1)
Re:Credibility? (Score:2)
Re:Credibility? (Score:1)
archive.org (Score:3, Interesting)
P.S. Are we losing information at a comparable rate to generation....?
interesting but... (Score:3, Interesting)
Phrase 3 mos 6 mos 12 mos. Total
buy low sell high 4700 5470 6200 7830
60% 70% 79% 100%
seems to demonstrate the opposite of the trend that he describes. Indeed, a current search on google [google.com] shows about 1,270,000 results (makes you wonder when he did his searches that the current number of results is so many orders of magnitude in difference). The methodology also fails to take in to account any growth in the size of the web, which could mask the effects of decay.
Re:interesting but... (Score:1)
Re:interesting but... (Score:1)
Re:interesting but... (Score:1)
The original seems higher than the most recent number... 'trend seems to be towards growth..'?
hmmm....
The Web is decaying (Score:5, Funny)
Yet another crippling bombshell hit the beleaguered web community when recently IDC confirmed that the web accounts for less than a fraction of 1 percent of all server usage. Coming on the heels of the latest Netcraft survey which plainly states that the web has lost more market share, this news serves to reinforce what we've known all along. The web is collapsing in complete disarray, as further exemplified by failing dead last [samag.com] in the recent Sys Admin comprehensive networking usage test.
You don't need to be a Kreskin [amdest.com] to predict the web's future. The hand writing is on the wall: the web faces a bleak future. In fact there won't be any future at all for the web because the web is decaying. Things are looking very bad for the web. As many of us are already aware, the web continues to lose market share. Red ink flows like a river of blood. Dot-coms are the most endangered of them all, having lost 93% of their core developers.
Let's keep to the facts and look at the numbers.
The web leader Theo states that there are 7000 users of the web. How many users of other protocols are there? Let's see. The number of the web versus other protocols posts on Usenet is roughly in ratio of 5 to 1. Therefore there are about 7000/5 = 1400 other protocols users. Web posts on Usenet are about half of the volume of other protocols posts. Therefore there are about 700 users of the web. A recent article put the web at about 80 percent of the HTTP market. Therefore there are (7000+1400+700)*4 = 36400 web users. This is consistent with the number of Usenet posts about the web.
Due to the troubles of Walnut Creek, abysmal sales and so on, the web went out of business and was taken over by Slashdot who sell another troubled web service. Now Slashdot is also dead, its corpse turned over to yet another charnel house.
All major surveys show that the web has steadily declined in market share. The web is very sick and its long term survival prospects are very dim. If the web is to survive at all it will be among hobbyist dabblers. The web continues to decay. Nothing short of a miracle could save it at this point in time. For all practical purposes, the web is dead.
Fact: the web is dead.
Re:The Web is decaying (Score:1)
The guy who posted this may have made a mistake. (Score:2, Informative)
Re:The guy who posted this may have made a mistake (Score:1)
Actually this post is getting old!!
Better article needed (Score:5, Interesting)
You seem to have some good ideas (Score:2)
So when can we expect to see your rigorous analysis? Or were you just bitching?
Exaclty who should be proving your thesis (Score:2)
BUT.... (Score:1)
But then again it is an interesting piece of trivia
we've lost the ability to rely on hyperlinks (Score:5, Insightful)
Re:we've lost the ability to rely on hyperlinks (Score:1, Funny)
Thought and mod_rewrite are the key (Score:5, Insightful)
Let's look at a few examples.
The URI to this page is http://slashdot.org/comments.pl?sid=31884&op=Repl
Well, for a start, that
Now let's look at an equivilent Kuro5hin [kuro5hin.org] URI.
http://www.kuro5hin.org/comments/2002/4/29/22137/
For a start, you can't tell what application or script is serving you the page, and you can't see what type of file it's linking to; both these things can and will change over time.
Second, there's a date embedded in there; you can see the developers, if they ever decide to change the meaning of '/comments', using that date as a reference; if the URI is before the change, they can map it onto the new schema or pass it onto legacy code.
Having the date in the URI is good because it allows you to determine when the link was issued, and map it onto any changes or pass it off to a legacy system as required.
Now let's take an apparantly good link on my now horribly out of date site, aagh.net [aagh.net].
http://www.aagh.net/php/style/ [aagh.net] links to an article on PHP coding style.
Certainly, hiding the fact that I'm using PHP to serve this document is good, and shortening the URI to remove the useless querystring is good (you can't see one? Good, that's the point), however, this URI may well stop working in a few weeks; I'm planning a redesign and the old schema may well not fit in well with it.
A short yyyymm in there could have made all the difference; a simple if check on the URI's issue date would keep it working.
The moral of the story: Think about your URI's when you're designing a site. Try to remove as much data as you can without painting yourself into a corner.
How do you account for domain changes? (Score:2)
The key to making links that don't rot is to design a URI schema that's both independent of any redesigns of your site and independent of any particular way of doing things.
You can't mod_rewrite a domain name that you have lost control over. If you have a popular site hosted on a university's server, and then you graduate, what do you do? If you put up a site, some Yakkestonian trademark holder takes it from you in WIPO court, and you're forced to go to Gandi.net to get a new domain, what do you do?
Re:How do you account for domain changes? (Score:2)
Nope, that's too bad. You can mod_rewrite a domain name you do have control over, though. You can also see if you can get the new owners to redirect to your new domain.
If, say, all your URI's start with a date, you ask the new owners to redirect any URI containing '/yyyy/mm/dd' less than the date you lost the domain to your new site. You may not get it all or even most of the time, but the option is still there.
Alas, this is one of those cases URN [ietf.org]'s would come in handy.
Re:Thought and mod_rewrite are the key (Score:2)
If I had mod points today, I would do it
Re:Thought and mod_rewrite are the key (Score:3, Informative)
The argument wasn't "query strings are bad, m'kay"; look at the URI and see what information's in there. Does
The URI's would work just as well using something like http://slashdot.org/stories/31884/comments/343680
> Do you know how to make k5's comments nested instead of threaded purely using the URI?
No. Actually, I wasn't really pointing out k5 as being the perfect example; Scoop actually tends to really suck in this respect (like setting the URI to '/' when you change comment modes). However, I might be tempted to ask you which URI is likely to live the longest, certainly back when SlashDot used to archive articles after a couple of weeks.
> The point is, wether or not it takes the optimum number of bytes isn't always the priority
I never once said the size of the URI was important. I said it contained a lot of extranious information that changed a lot while meaning little (i.e. the URI's changed from the dynamic query string to an
> in the case of
What's easier for the "savvy" user? A URI that will work for the rest of SlashDot's life, or one that'll last until the story is archived, or the underlying architecture changes, and which contains a lot of randomly ordered and mostly meaningless information?
A well designed URI scheme will actually give the savvy user a lot more control; say, you include the date of an article, ala http://slashdot.org/stories/2002/05/30/; you can imagine going to such a URI and getting all of the stories on that day, month or year, and instantly being able to identify how old a linked to article is. You can also imagine an archived URI and a live, dynamic URI both using the same schema.
You can also imagine giving a URI of an interesting article to a friend without first having to decode the query string; just strip off anything after
Note: This applies to any site, those particular SlashDot and k5 URI's were just examples.
Re:we've lost the ability to rely on hyperlinks (Score:2)
filenames: people change languages (php, perl, asp, etc), site layouts, functionality.
filenames: intentionally changed to prevent deep linking (heh heh)
When I change a URI it is usually because I'm changing the logical structure of that program. However I also usually check the referrer logs, and if there has been an outside referral then I will put in a redirect for the old file, and contact the site that had the link to ask them to change it.
There is no excuse for having broken links on your own site though, though it does happen to the best of us
Travis
Re:we've lost the ability to rely on hyperlinks (Score:2)
And this costs me nothing but "oh yeah, must remember to upload index.html as well as index.htm".
I expect that my sites will be just as valid 20 years from now, assuming Earthlink is still in business and still hosting it. (Yeah, I should get my own domain names, but..)
If you have dynamic content, your needs might differ. But for informational sites, change for the sake of change is usually a Bad Thing.
Stop Web Decay Today (Score:4, Funny)
Re:Stop Web Decay Today (Score:1)
this 'study' suggests to me that there is room for real scientific investigation into the nature of massively webbed information. and google very likely provides a useful tool in the information-scientist's investigative arsenal.
Self-fulfilling prophecy? (Score:1, Funny)
Temporarily Unavailable
The Angelfire site you are trying to reach has been temporarily suspended due to excessive bandwidth consumption.
The site will be available again in approximately 2 hours!
Study? (Score:4, Insightful)
I appreciate the topic very much, but some more material on it is needed. This study wouldn't be complete enough even for high-school homework...
And look at his homepage (just remove the last part of the url). The most pages are more than two years old... that's decay!
Seriously speaking, just look for a few more sources before you accept a story.
Study claims ?? (Score:1, Insightful)
The guy that submited this story is the guy that did the study.
Google/CowboyNeal Study (Score:5, Funny)
Google gives us the following interesting results:
3,840,000 [google.com] sites contain the word Cheese.
1,640 [google.ch] sites contain the words CowboyNeal and Cheese.
Therefore, 4.27083333333333333333333333333e-2% of cheese related sites contain a reference to CowboyNeal.
As cheese is a randomly chosen word with no special connection to CowboyNeal it is reasonable to assume that 4.27083333333333333333333333333e-2% of all sites contain a reference to The Cowboy (Assuming the number of sites dedicated to CowboyNeal equals the number dedicated to ignoring him).
So there we have it. The web is 99.957291666666666666666666666667% CowboyNeal free.
I said the results were "precise", not "accurate".
Re:Google/CowboyNeal Study (Score:1)
Re:Google/CowboyNeal Study (Score:4, Funny)
S
Re:Google/CowboyNeal Study (Score:2)
Your assumptions seem reasonable. I'm a reasonable man, and I don't know nothing about no CowboyNeal and his cheese fetish. But when you get to an absurd result like
Ergo, CowboyNeal has a "special connection" to cheese. Quod Erat Doodah.
major source of web decay: (Score:1)
Not that accurate (Score:2)
Jakob Nielsen: Web Pages Must Live Forever (Score:3, Interesting)
See this letter I wrote to TimBL (Score:2)
Once you have put a page on the Web, you need to keep it there indefinitely.
How is this possible if you happen to lose control of the domain? I wrote a letter to Tim Berners-Lee about this issue.
One More Reason Why The Web Sucks (Score:1)
The Angelfire site you are trying to reach has been temporarily suspended due to excessive bandwidth consumption."
Imagine that you were renting a building and running a business - a retail store. One day, the owner of the bulding comes in and padlocks the doors and says "Sorry, you can't re-open till the first of the month - too many people have come into your store".
What stupidity.
Re:One More Reason Why The Web Sucks (Score:1)
If you're using someone else's building (for no cost, mind), this person certainly has the right to kick you out if he feels "too many people have come by".
Needs a broader range (Score:1)
If he really wanted a large search he should have tried "porn".....
Heh.. Talk about web decay. (Score:5, Funny)
Looks like 100% of the link mentioned in this article decayed in a little under 5 minutes!
Cheers,
Re:Heh.. Talk about web decay. (Score:1)
Google "pages found" data (Score:3, Interesting)
While the numbers clearly aren't totally random, they are very fragile indeed. Some people have had a change of two orders of magnitude, within a week. And in these cases, there have usually been no real world events that could explain such a change. I guess the google page hits numbers depend as much on the internal google structure, as on the number of actual pages on the web.
So I doubt google page hits statistics is a useful research tool. Nonetheless, it can be fun. Here are some google hall of fame lists:
Correct links (Score:3, Interesting)
average and real life persistance of documents (Score:2)
not that it was an interesting document - just a little paper about nothing important. But still, it's out there.
My thoughts? I think that as long as a website can be "saved" in some form, its content will be available in other forms for a long amount of time.
this should make people think, especially those who put copyrights on their webpages, or don't want some information to spread around.
could we say that information want to be free as long as it's downloadable?
hmm..
I conclude there is no decay ! (Score:1)
Sex
Warez
mp3
I have discovered that amazingly, my results differ substantially !
In conclusion, then, it seems that content is ultimately always fresh and there is no indication of decay !
Wide jump from findings to conclusion (Score:5, Interesting)
What scares me here is the conclusion that web sites need to change their content 60% every 3 months. This is not freshness, this is reorganizing to re-organize. If you are considering doing this, you had better seriously re-consider your future. Its an interesting study but a good meme doesn't die simply because the catch-phrases are tired.
At faculty meetings at our school I sit with a bingo card. On it are a series of catch-phrases. We listen for the catch-phrases and shout out when we have finished our cards. B***SH*T is the game and to reduce your content to a series of reorganized catch-phrases is like having a marketing guy develop foreign policy.
Anyone willing to write the perl module that searches for the latest catch-phrases and inserts them randomly into your web content. Yeesh!
Google measurements--varied and fun! (Score:2, Funny)
Decay (Score:1)
The Slashdotted Decay Rate (Score:1)
:)
Google Study in Another Place (Score:5, Informative)
http://helen.lifeseller.com/webdecay.html [lifeseller.com]
I've also included a link to the raw data I used.
Local storage of information that I want to keep (Score:1)
For years, whenever I've found an article that I've liked, or data that I thought would be useful later on, I've always either saved the
Ryan
guidelines for content change? (Score:2)
This seems so totally- "if everyone else is
jumping off the Brooklyn Bridge, then we
should to" by itself that it discredits what
sliver of credibility the article had. Using
a web-wide average as a guideline for what
a particular web site "should do" is
meaningless. Web sites should present timely,
appropriate information that is useful to
those who visit. Some sites deal with
material that changes frequently (stock quotes
and sports sites should be presumably updated
regularly) and some sites deal with material
that does not change frequently (no need to
redo your tech support documents for long-
out of production products every week.)
This notion of `freshness' is ill-defined,
poorly measured and of dubious value.
Irony! (Score:2)
Re:Applying statistics meaningfully (Score:1)