How Many Google Machines, Really? 476
BoneThugND writes "I found this article on TNL.NET. It takes information from the S-1 Filing to reverse engineer how many machines Google has (hint: a lot more than 10,000).
'According to calculations by the IEE, in a paper about the Google cluster, a rack with 88 dual-CPU machines used to cost about $278,000. If you divide the $250 million figure from the S-1 filing by $278,000, you end up with a bit over 899 racks. Assuming that each rack holds 88 machines, you end up with 79,000 machines.'" An anonymous source claims
over 100,000.
Nice Rack! (Score:4, Funny)
What is that as a percentage ... (Score:3, Interesting)
* of servers in the USA
* of servers running Linux
Re:What is that as a percentage ... (Score:5, Funny)
Re:What is that as a percentage ... (Score:5, Funny)
Re:What is that as a percentage ... (Score:4, Funny)
Re:What is that as a percentage ... (Score:3, Funny)
0.0001%
* of servers in the USA
0.00000045%
So there are 222 times as many servers in the United States as in the entire world, and there are 1.96e13 servers in the USA. Too bad there isn't a "-1 bad arithmetic" moderation....
$278k ?? (Score:5, Insightful)
Re:$278k ?? (Score:4, Informative)
Re:$278k ?? (Score:3, Flamebait)
imagine what google could do with that kind of shit
Re:$278k ?? (Score:3, Insightful)
What's cheaper... buying a round robin DNS router (hardware) or coding your client to try the next web server in it's list (software). Now, multiply that savings for every customer you sell to.
The problem is finding someone who knows how to do that robustly and reliably. Most places have troubling finding developers whose programs don't crash every 15 minutes. This sort of thing is a little more adva
Re:$278k ?? (Score:5, Insightful)
The difference is that if Google loses track of a few pages due to node failure it's no big deal because a) they don't guarantee to index every page on the web anyway and b) the chances are that page will be spidered again in the near future - and it may not even still exist anyway.
Your bank, on the other hand, can't just "lose" a few transactions here and there. FedEx can't just lose a few packages there and there. Sure they occasionally physically lose one, but they never lose the information that at one point, they did have it. Your phone company can't just lose a few calls you made and not bill you for them. Your hospital can't just lose a few CAT scans and think oh well, he'll be in for another scan eventually.
Now, I'm not saying that Google's technique isn't clever - I'm saying that it can't really be generalized to other applications. And that's why very smart people - and big corporations can afford to hire very smart people - keep on buying Sun and IBM kit by the boatload.
Interesting list... (Score:5, Funny)
Wait, what's wrong with that one?
Re:$278k ?? (Score:5, Interesting)
You've never worked in a medical field have you? You'd think that that would be a big deal and in theory data integrity is a very high priority but in reality...
I used to work as the IT Manager for a diagnostic imaging and cancer treatment center (and still do contract work with them because my replacement is kind of a noob) While loosing studies isn't exactly a "no big deal" situation it's still far more common than patients will ever realize. The server that stores and processes all of the digital images from the scanning equipment is a single CPU home rolled P4 using some shitty onboard IDE raid controller (doesn't even do RAID5!) running Windows 2K. The most money I could get for setting up a backup solution was the $200 an external firewire drive cost. Somehow we never managed to loose a study once it reached my network in the 9 months I worked there but I know three or four were deleted from the cameras themselves before being sent properly so whoops it's gone, gotta reschedule (and bill their insurance or Medicare again!) Two weeks ago one of the drives in that 0+1 array failed and despite my pleadings they still haven't ordered a replacement yet...
Now it's tempting to think that this place is just a special case of cheapness and sloppiness but from talking to the diagnostic techs (the people that operate the cameras) that's not so. That clinic is a little worse than average in terms of loosing patient information but by no means the worst some of them at seen/heard of/worked at in their careers. It's worse in general at small facilities but even large hospitals often suffer from the same unprofessionalism.
Your bank and the phone company keep much better track of your calls or your ATM transactions than most hospitals do with your CT or MRI scans...
Redundancy (Score:4, Interesting)
Some of the reasons these techniques aren't used in enterprise computing:
Since I've seen it up close a few times, I can say that the standard "enterprise way" (Oracle/Sun/EMC) delivers very poor bang for the buck. If Google wanted to, they could deliver a modified GFS with any desired level of reliability by increasing the redundancy. And even after that bloating, it would still deliver greater bang for the buck than the conventional solutions.
Re:$278k ?? (Score:5, Informative)
Re:$278k ?? (Score:5, Informative)
Doesn't work like that, kid. A CPU on a high-end Sun fails, and the system will keep on running. You can swap the CPU out and replace it with a new one, the system will simply pick it up, assign threads to it, and keep on running. Had a couple of CPUs fail a little while ago... the first we users noticed of it was that the application slowed down slightly. Sysadmin just said yeah, I know, I'll replace 'em when the parts arrive this afternoon. Cool, we said. No data lost, no need to shut down or even restart our app. 'Course you gotta architect your app to deal with that - like don't have just one thread that does a crucial task, 'cos there's a chance that might be on the CPU that fails. But still, it's no big deal.
Re:$278k ?? (Score:5, Informative)
And hey, if you want to mix and match CPU types (uSparc 2 and 3, etc), speeds, etc, no problem either. So if you wanna upgrade your server's CPUs, there will be zero downtime, you just do it a board at a time (board = 2 or 4 CPUs).
lego? (Score:3, Insightful)
Re:lego? (Score:3, Interesting)
Re:$278k ?? (Score:5, Interesting)
Re:$278k ?? (Score:5, Insightful)
Re:Not to sound like your Mom [or Señor Ashcr (Score:4, Funny)
Re:Not to sound like your Mom [or Señor Ashcr (Score:3, Informative)
Of course, most of it is more for show than practicality. I mean, they have hand scanners on every single cage. Definitely a little bit excessive
Acquisition (Score:5, Insightful)
The cost of acquiring the machine is a fraction of the cost of owning it.
And lets not forget the overhead of 2 networks per machine and all the patch panels, wiring, switches. Toss in console management (which may not be on all machines at all time), monitoring and management of said machines. Oh, and one really tired guy running around.
Disks are going to fail at a rate of several hundred or thousand PER DAY, just statistically. (along with power supplies etc)
Toss in that in three years, ALL of those machines are obsolete.
That's huge.
I've got ~300 racks in a half full data center upstairs from me. All network cables run to a room below it to patch panels. Around 50% the size of the DC is cable management. Next to that is a room FILLED with chest high batteries - these are used during outages until the generators need to be kicked on. And a NOC takes up about 1/5th the space of the DC (monitoring systems worldwide, but it's got seating for maybe 40 people - tight and usually filled with 10 folks, but in a crunch we live up there).
So that $3159 is only a bit of it. And in 3 years, all those machines will likely be replaced for whatever $3k buys then. That's about to be a 2 CPU Athlon64 box. If Sun can pull a rabbit out of its ass, we'll have 8 and 16CPU Athlon64 boxes. At least with that, some of the CPUs can talk to each other really really really fast.
Re:Acquisition (Score:5, Informative)
that's a little over the top big guy. i've worked at a 10,000 node corp doing desktop support. We lost ONE disk perhaps a week....if that much. We often went several weeks with no disks lost.
even if you factor in multiple drives per server, say TWO (because they are servers not desktops)
Interpolate for 100,000, that's a max of 20 disks per week...on the high end.
Can you imagine (Score:4, Funny)
Re:Can you imagine (Score:5, Funny)
Google, will you marry me? (Score:5, Funny)
2) google is worth so much money
3) google has a huge rack!!
Re:Google, will you marry me? (Score:5, Funny)
Love,
Yahoo.
Re:Google, will you marry me? (Score:4, Funny)
Re:Google, will you marry me? (Score:3, Funny)
IPO changes things (Score:5, Interesting)
Why do we care? (Score:4, Funny)
Re:Why do we care? (Score:5, Funny)
I'm more interested.. (Score:5, Funny)
Better yet, open up a nursery (plant type) next door , build a green house, and piple 25% of the heat to it. Have you guys see the price of trees lately? Google could make a killing with the "recycling" plant.
Re:I'm more interested.. (Score:4, Funny)
I guess that's one way to solve the homeless problem.
Not unexpected... (Score:5, Insightful)
At $699 per CPU (Score:5, Funny)
Re:At $699 per CPU (Score:4, Funny)
Re:At $699 per CPU (Score:3, Informative)
What a waste (Score:3, Funny)
I hang around too many old-timer mainframe geeks. MVS forever!!! and such.
Re:What a waste (Score:5, Interesting)
Sunny Dubey
Re:What a waste (Score:5, Informative)
Mainframes are optimized for batch processing. Interactive queries do not take full advantage of their vaunted I/O capacity.
Moreover, while a mainframe may be a good way to host a single copy of a database that must remain internally consistent, that's not the problem Google is solving. It's trivial for them to run their search service off of thousands of replicated copies of the Internet index. Even the largest mainframe's storage I/O would be orders of magnitude smaller than the massively parallel I/O operations done by these thousands of PCs. Google has no reason to funnel all of the independent search queries into a single machine, so they shouln't buy a system architecture designed to do that.
Assumptions? (Score:5, Interesting)
Um, don't you think if you were buying 899 racks you might actually, you know, negotiate for a better price?
This isn't the only assumption in your analysis, and the problems with them will be compounded. What's the point of this, really?
Re:Assumptions? (Score:5, Funny)
Re:Assumptions? (Score:5, Insightful)
Maybe just me... (Score:4, Insightful)
wait (Score:5, Insightful)
It's gotta be more than that.
All that power (Score:5, Funny)
Really? (Score:5, Funny)
This is actually useful (Score:3, Interesting)
where do you go to buy 80,000 hard drives?
Re:This is actually useful (Score:5, Insightful)
You don't; their Sales Director comes to you...
Re:This is actually useful (Score:5, Funny)
88 machines per rack? hardly. (Score:3, Interesting)
Re:88 machines per rack? hardly. (Score:5, Informative)
If Google is innovating in this area, it could either be on price or in density.
Re:88 machines per rack? hardly. (Score:3, Interesting)
Power (Score:3, Funny)
Google hosting (Score:5, Interesting)
Re:Google hosting (Score:5, Insightful)
This is a good thing for Google, but not for the world as a whole.
Re:Google hosting (Score:5, Interesting)
Re:Google hosting (Score:4, Interesting)
Someone call the FBI (Score:5, Funny)
I have seen the light (Score:3, Informative)
they fit about 100 or so 1u's on each side of the rack, there double sided cabinets that look like refrigerators. there seperated in the center by noname brand switches and they have castor wheels on the bottoms of them. google can at the drop of a dime roll there machines out of a datacenter onto there 16 wheeler, move, unload and plug into a new data center in less than a days time.
Re:Corrected version - Re:I have seen the light (Score:5, Funny)
Obviously.
Re:Corrected version - Re:I have seen the light (Score:4, Funny)
Obviously.
I'm not a grammer/spelling Nazi.
Obviously.
Makes Perfect Since (Score:3, Insightful)
It would only make sense that the server count would now be in the ballpart of what is mentioned here.
Google hasn't been standing still, and I've heard the "Google has 10k servers" for 1-2 years now.
-Pete
15 Megawatts (Score:5, Interesting)
...assuming 200W per server, which is probably low, but probably compensates for 79,000 being most likely an overestimate. However, that doesn't even begin to account for the energy used to keep the stuff cool.
Anyone know how many trees per second that would be? Conversion to clubbed-baby-seals-per-sec optional.
Re:15 Megawatts (Score:5, Funny)
We're on to you, Google!
Re:15 Megawatts (Score:5, Funny)
Hm... Google seems decidedly male to me.
1) Answers questions rapidly without offering any description of how the answer was derived? Check.
2) Works in short, fast bursts of energy and then tells you proudly it only took them .009 seconds? Check
3) Has an inability to accessorize his appearance? Check.
4) Returns 82,200,000 results when asked about porn? Check and match!
Re:15 Megawatts (Score:3, Funny)
The servers could be powered by 15 Megahamsters on treadmills (@ 1 watt/hamster). But that would require sufficient management to motivate the hamsters with the threat of off-shoring their jobs.
Heat (Score:5, Informative)
I hope they have good ventilation...
Re:Heat (Score:5, Funny)
<sound of door slamming>
<sound of car engine starting>
<sound of tires squeeling away>
Re:Heat (Score:5, Funny)
"I'm meeeelllllttttiinnnnnggggg"
Bob.
Re:Heat (Score:5, Informative)
Ulrik
Re:Heat (Score:3, Funny)
Now, where's my Delorean?
Cheers!
SCO (Score:3, Insightful)
The interesting thing is, that if SCO really has MS backing and MS is pulling strings, then I would think that MS would want SCO to persue google to tie them up for awhile.
hardcore (Score:5, Funny)
42.
Why reverse engineer... (Score:5, Informative)
Google has 3 sites (two west coast, one east)
Each site connected with 1 OC48
Each OC48 hooks up to 2 Foundry BigIron 8000
80 Pc's per rack * 40 racks(at an example site)
= 3200 PC's.
A google site is not a homogenous set of PC's instead there are different types of PC's that are being upgraded on different cycles based on the price/performance ratio.
If you want more info get the patterson hennessy book that I mentioned. Not the other version they sell. This one rocks way harder. You get to learn fun things like Tomosulo's algorithm.
If I am violating any copy rights feel free to remove this post.
inside information (Score:5, Interesting)
Interesting People 2004/05 [interesting-people.org]:
I know for a FACT they passed 100,000 last November. One thing the Louis calculation may have missed is Google's obsession with low cost. For example read the company's technical white paper on the Google file system. It was designed so that Google could purchase the cheapest disks possible, expecting them to have a high failure rate. What happens when you factor cost obsession into his equation?
Re:inside information (Score:4, Informative)
Ulrik
Absolutely Beautiful (Score:5, Insightful)
Thanks Google!
You're not factoring in Google's culture (Score:4, Interesting)
False advertising. (Score:5, Funny)
Environmental impact: power to 68,000 homes (Score:3, Funny)
The 1,100 Apple cluster at Virginia tech uses 3 megawatts, sufficient to power 1,500 Virgina homes
http://www.research.vt.edu/resmag/2004resmag/HowX
Yes, it is true: every time you hit Google, you are polluting the Earth.
Re:Environmental impact: power to 68,000 homes (Score:3, Interesting)
Whereas Slashdot uses nothing but solar power.
Re:Environmental impact: power to 68,000 homes (Score:5, Insightful)
Scary... DDOS? (Score:3, Interesting)
Someone mentioned that they have enough bandwidth/processing power to saturate a T1000 line. Scary...
No (Score:5, Informative)
But his low end number are Wrong... (Score:3, Interesting)
Server pricing (Score:5, Informative)
Every article I've read about Google's servers says they use "commodity" parts, which means they buy pretty much the same stuff we buy. They also indicate that they use as much memory as possible, and don't use hard drives, or use the drives as little as possible. From my interview with Google, they asked quite a few questions about RAID0, RAID1 (and combinations of those), I'd believe they stick in two drives to ensure data doesn't get lost due to power outages.
We get good name brand parts wholesale, which I'd expect is what they do too. So, assuming 1u Asus, Tyan, or SuperMicro machines stuffed full of memory, with hard drives big enough to hold the OS plus an image of whatever they store in memory (ramdrives?), they'd require at most 3Gb (OS) + 4Gb (ramdrive backup). I don't recall seeing dual CPU's, but we'll go with that assumption.
The nice base machine we had settled on for quite a while was the Asus 1400r, which consisted of dual 1.4Ghz PIII's, 2Gb RAM, and 20Gb and 200Gb hard drives. Our cost was roughly $1500. They'd lower the drive cost, but incrase the memory cost, so they'd probably cost about $1700, but I'm sure Google got better pricing, buying the quantity they were getting.
The count of 88 machines per rack is a bit high. You get 80u's per standard rack, but you can't stuff it full of machines, unless you get very creative. I'd suspect they have 2 switches, and a few power management units per rack. The APC's we use take 8 machines per unit, and are 1u tall. There are other power management units, that don't take up rack space, which they may be using, but only the folks at Google really know.
Assuming the maximum density, and equipment that was available as "commodity" equipment at the time, they'd have 2 Cisco 2948's and 78 servers per rack.
$1700 * 78 (servers)
+
$3000 * 2 (switches)
+
$1000 (power management)
--------
$139,600 per rack (78 servers)
Lets not forget core networking equipment. That's worth a few bucks.
Each set of 39 servers would probably be connected to their routers via GigE fiber (I couldn't imageine them using 100baseT for this) Right now we're guestimating 1700 racks. They have locations in 3 cities, so we'll assume they have at least 9 routers. They'd probably use Cisco 12000's, or something along that line. Checking eBay, you can get a nice Cisco 12008 for just $27,000, but that's the smaller one. I've toured a few places who had them, and pointed at them citing them to be just over $1,000,000.
So....
$250,000,000 (ttl expenses)
- $ 9,000,000 (routers)
------
$241,000,000
/ $ 139,600
------
1726 racks
* 78 (machines per rack)
------
134,682 machines
Google has a couple thousand employees, but we've found that our servers make *VERY* nice workstations too.
I believe this to be a more fair estimate, than the story gave. They're quoting pricing for a nice fast *CURRENT* machine, but Google has said before that they buy commodity machines. They do like we do. We buy cheap (relatively) and lots of them, just like Google does. We didn't pattern ourselves after Google, we made this decision long before Google even existed.
When *WE* decided to go this router, we looked at many options. The "provider" we had, before we went on our own, leasing space and bandwidth directly from Tier 1 providers, opted for the monolythic sy
Re:Server pricing (Score:4, Insightful)
Which is why they have no problems finding space for GMail - you can't buy full size drives as small as 7Gb anymore, so they already have countless Tbs of unused drive space in their racks.
I think they include infrastructure & air cool (Score:3, Interesting)
Re:Pretty Broad (Score:5, Insightful)
Re:Pretty Broad (Score:4, Funny)
Your wife has slept with 80 other men, or was it 200?
Either way, it's not good for you.
Re:Which brings up an interesting question... (Score:4, Interesting)
This is how it should be, since knowing the size of Google's hardware capacity is a very, very strategic bit of information, and the kind of thing that would allow Yahoo/MSN/whoever to get a feel for how much capital would be necessary to duplicate or improve upon it.
Re:Ask (Score:3, Funny)
Say, Jeeves old boy: how many servers does Google have?
Jeeves: Piss off!
Re:The things you could do with that... (Score:3, Funny)
What about building a really big search engine?
To build a really big search engine, your going to need some serious distributed computing, and a big database! Hey wait a minute thats what they are doing!
Re:They don't need 40K machines! (Score:3, Insightful)
Re:Nobody has 88 systems in a rack (Score:5, Interesting)
Re:Come on! Does it really matter? (Score:5, Funny)
Well, that depends on what sort of time portal they use. Now, a T1000 would probably be saturated by a time portal following the Terminator rules: one way only. But Google seems to favour Back to the Future rules, as shown by number of hits:
13,500,000 for back to the future
3,460,000 for terminator
This would make saturating a T1000 a lot easier, since you could saturate it while travelling back in time yourself, or maybe even while standing still in time. This would make Google's bandwidth infinite, as a measly T1000 would stand still. Unless it was using its own time portal to travel back in time to destroy Google, but that would create a paradox, since, as we all know, Google will become Skynet, which will create the T1000 in the first place.
What I'm trying to say is: I don't know, but I'm sure Google could do it.