I’ve got a new job, and my employer has assigned me to back up the internet. Just the parts that are publically accessible. I have 1gb backup media. How many media will I need?
A warehouse full of media. And eleventeen lifetimes in which to perform your duties. Have fun.
Did you possibly mean Intranet or is your boss a mental case?
Amp
February 5, 2006, 5:48am
4
Are they letting you back-up porn too? That is such a sweet job.
you might like to check out this site .
Seven
February 5, 2006, 6:17am
6
After you’re done backing up the Internet, you can safely shut it down here
By the way, this is the end of the internet
Folly
February 5, 2006, 6:25am
7
Tell your boss you have to to download the newest version of the internet first.
Ok, ok. I’m thinking this is Squeegee ’s way of asking how many gigabytes the publicly accessible portions of the internet are. Any estimates?
Folly:
Tell your boss you have to to download the newest version of the internet first.
Ok, ok. I’m thinking this is Squeegee ’s way of asking how many gigabytes the publicly accessible portions of the internet are. Any estimates?
Did you ever watch Cosmos on PBS, I think Carl Sagan’s Catch Phrase of “Billions and Billions” is about right to describe the size.
Jim
Gigabytes???
All of the publicly available internet hypermedia?
More like Terabytes, many thousands of terabytes.
Terabytes? More like petabytes. Or exabytes. Not sure if it reaches zettabytes, much less yottabytes yet, tho.
I guess you work at google, huh?
Q.E.D
February 5, 2006, 6:57am
12
Some (slightly outdated) information here and here .
The “surface” Web consists of approximately 2.5 billion documents [1 and 5], up from 1 billion pages at the beginning of the year [3], with a rate of growth of 7.3 million pages per day [1]. Estimates of the average “surface” page size vary in the range from 10 kbytes [1] per page to 20 kbytes per page [4]. So, the total amount of information on the “surface” Web varies somewhere from 25 to 50 terabytes of information [HTML-included basis]. If we want to obtain a figure for textual information, we would use a factor of 0.4 [4], which leads to an estimate of 10 to 20 terabytes of textual content. At 7.3 million new pages added every day, the rate of growth is [taking an average estimate] 0.1 terabytes of new information [HTML-included] per day.
If we take into account all web-accessible information, such as web-connected databases, dynamic pages, intranet sites, etc., collectively known as “deep” Web, there are 550 billion web-connected documents, with an average page size of 14 kbytes, and 95% of this information is publicly accessible [2]. If we were to store this information in one place, we would need 7,500 terabytes of storage, which is 150 times more storage than we would need for the entire “surface” Web, even taking the highest estimate of 50 terabytes. 56% of this information is the actual content [HTML excluded], which gives us an estimate of 4,200 terabytes of high-quality data. Two of the largest “deep” web sites - National Climatic Data Center and NASA databases - contain 585 terabytes of information, which is 7.8% of the “deep” web. And 60 of the largest web sites contain 750 terabytes of information, which is 10% of the “deep” web.
When your done, can I get a copy?
Probably all the OP is referring to is backing up the company’s publicly accessible web site.
In this case, I’d suggest getting an FTP client, logging on the server and dragging the folders from the server to your desktop PC. How much media you’ll need depends on how extensive the internet presence is, of course.
Precisely. Sorry if I confused anyone.
OK, so I’d need 7500 of my media. Or 938 double-sided DVDs. Or perhaps considerably less, since HTML would compress quite well.
Better make a hardcopy just in case.
Just
February 5, 2006, 3:13pm
19
Shouldn’t that be 7,500,000 of your media?
Damn, you’re right. a Terabyte is a 1000 gigabytes, so I’d need 7.5mil media to back up 7500 terabytes.