Google Data Centers

Google data centers are the large data center facilities Google uses to provide their services, which combine large drives, computer nodes organized in aisles of racks, internal and external networking, environmental controls (mainly cooling and humidification control), and operations software (especially as concerns load balancing and fault tolerance).

Google Data Centers
Former Google data center in Eemshaven, Netherlands
External videos
YouTube logo
YouTube
video icon Google Data Center 360° Tour

There is no official data on how many servers are in Google data centers, but Gartner estimated in a July 2016 report that Google at the time had 2.5 million servers. This number is changing as the company expands capacity and refreshes its hardware.

Locations

The locations of Google's various data centers by continent are as follows:

Continent Location Geo Products Location Cloud Location Timeline Description
North America Arcola (VA), USA 38°56′35.99″N 77°31′27.61″W / 38.9433306°N 77.5243361°W / 38.9433306; -77.5243361 Loudoun County N. Virginia (us-east4) 2017 - announced
North America Atlanta (GA), USA 33°44′59.04″N 84°35′5.33″W / 33.7497333°N 84.5848139°W / 33.7497333; -84.5848139 Douglas County - 2003 - launched 350 employees
South America Cerrillos, Santiago, Chile 33°31′14″S 70°43′18″W / 33.520515°S 70.721695°W / -33.520515; -70.721695 - Santiago (southamerica-west1) 2020 - announced

2021 - launched

Asia Changhua County, Taiwan 24°08′18.6″N 120°25′32.6″E / 24.138500°N 120.425722°E / 24.138500; 120.425722 Changhua County Taiwan

(asia-east1)

2011 - announced

2013 - launched

60 employees
North America Clarksville (TN), USA 36°37′16″N 87°15′47″W / 36.6211599°N 87.2630735°W / 36.6211599; -87.2630735 Montgomery County - 2015 - announced
North America Columbus (OH), USA - Columbus (us-east5) 2022 - launched
North America Council Bluffs (IA), USA 41°13′17.7″N 95°51′49.92″W / 41.221583°N 95.8638667°W / 41.221583; -95.8638667 Council Bluffs 2007 - announced

2009 - completed first phase completed

2012 and 2015 - expanded

130 employees
North America Council Bluffs (IA), USA 41°10′06″N 95°47′46″W / 41.1682533°N 95.7961246°W / 41.1682533; -95.7961246 Iowa (us-central1)
Asia Delhi, India - Delhi (asia-south2) 2020 - announced

2021 - launched

Middle East Doha, Qatar - Doha (me-central1) 2023 - launched
Europe Dublin, Ireland 53°19′12.39″N 6°26′31.43″W / 53.3201083°N 6.4420639°W / 53.3201083; -6.4420639 Dublin - 2011 - announced

2012 - launched

150 employees
Europe Eemshaven, Netherlands 53°25′31″N 6°51′44″E / 53.4252171°N 6.8622574°E / 53.4252171; 6.8622574 Eemshaven 2014 - announced

2016 - launched

2018, 2019 - expansion

200 employees
Europe Frankfurt, Germany 50°07′21″N 8°58′27″E / 50.1226299°N 8.9741678°E / 50.1226299; 8.9741678 - Frankfurt (europe-west3) 2022 - expanded
Europe Fredericia, Denmark 55°33′29.5″N 9°39′20.8″E / 55.558194°N 9.655778°E / 55.558194; 9.655778 Fredericia - 2018 - announced

2020 - launched

€600M building costs
Europe Ghlin, Hainaut, Belgium 50°28′09.6″N 3°51′55.7″E / 50.469333°N 3.865472°E / 50.469333; 3.865472 Saint-Ghislain Belgium (europe-west1) 2007 - announced

2010 - launched

12 employees
Europe Hamina, Finland 60°32′11.68″N 27°7′1.21″E / 60.5365778°N 27.1170028°E / 60.5365778; 27.1170028 Hamina Finland

(europe-north1)

2009 - announced

2011 - first phase completed

2022 - expansion

6 buildings, 400 employees
North America Henderson (NV), USA 36°03′20″N 115°00′37″W / 36.0556248°N 115.0102258°W / 36.0556248; -115.0102258 Henderson Las Vegas (us-west4) 2019 - announced

2020 - launched

64-acres; $1.2B building costs
Asia Hong Kong, Hong Kong - Hong Kong (asia-east2) 2017 - announced

2018 - launched

Asia Inzai, Japan 35°49′04″N 140°07′57″E / 35.8177804°N 140.1323916°E / 35.8177804; 140.1323916 Inzai - 2023 - launched
Asia Jakarta, Indonesia - Jakarta (asia-southeast2) 2020 - launched
Asia Koto-Ku, Tokyo, Japan - Tokyo

(asia-northeast1)

2016 - launched
North America Leesburg (VA), USA 39°3′6.38″N 77°32′20.38″W / 39.0517722°N 77.5389944°W / 39.0517722; -77.5389944 Loudoun County N. Virginia (us-east4) 2017 - announced
North America Lenoir (NC), USA 35°53′54.78″N 81°32′50.58″W / 35.8985500°N 81.5473833°W / 35.8985500; -81.5473833 Lenoir - 2007 - announced

2009 - launched

over 110 employees
Asia Lok Yang Way, Pioneer, Singapore 1°19′26″N 103°41′36″E / 1.3239859°N 103.693253°E / 1.3239859; 103.693253 Singapore Singapore (asia-southeast1) 2022 - launched
Europe London, UK - London

(europe-west2)

2017 - launched
North America Los Angeles (CA), USA - Los Angeles (us-west2)
Europe Madrid, Spain 40°31′10″N 3°20′27″W / 40.5195333°N 3.3409366°W / 40.5195333; -3.3409366 - Madrid (europe-southwest1) 2022 - launched
Pacific Melbourne, Australia - Melbourne

(australia-southeast2)

2021 - launched
Europe Middenmeer, Noord-Holland, The Netherlands 52°47′24″N 5°01′45″E / 52.7901053°N 5.0292193°E / 52.7901053; 5.0292193 Middenmeer Netherlands (europe-west4) 2019 - announced
North America Midlothian (TX), USA 32°26′35″N 97°03′44″W / 32.443170°N 97.062324°W / 32.443170; -97.062324 Midlothian Dallas (us-south1) 2019 - announced

2022 - launched

375-acres; $600M building costs
Europe Milan, Italy - Milan (europe-west8) 2022 - launched
North America Moncks Corner (SC), USA 33°03′50.8″N 80°02′36.1″W / 33.064111°N 80.043361°W / 33.064111; -80.043361 Berkeley County South Carolina (us-east1) 2007 - launched

2013 - expanded

150 employees
North America Montreal, Quebec, Canada - Montréal (northamerica-northeast1) 2018 - launched 62.4-hectares; $600M building costs
Asia Mumbai, India - Mumbai (asia-south1) 2017 - launched
North America New Albany (OH), USA 40°03′41″N 82°45′31″W / 40.0613352°N 82.7586278°W / 40.0613352; -82.7586278 New Albany - 2019 - announced 400-acres; $600M building costs
Asia Osaka, Japan - Osaka

(asia-northeast2)

2019 - launched
South America Osasco, São Paulo, Brazil - São Paulo (southamerica-east1) 2017 - launched
North America Papillion (NE), USA 41°08′00″N 96°08′39″W / 41.1332915°N 96.144178°W / 41.1332915; -96.144178 Papillion - 2019 - announced 275-acres; $600M building costs
Europe Paris, France - Paris (europe-west9) 2022 - launched
North America Pryor Creek (OK), USA 36°14′28.1″N 95°19′48.22″W / 36.241139°N 95.3300611°W / 36.241139; -95.3300611 Mayes County - 2007 - announced

2012 - expanded

over 400 employees, land at MidAmerica Industrial Park
South America Quilicura, Santiago, Chile 33°21′30.5″S 70°41′50.4″W / 33.358472°S 70.697333°W / -33.358472; -70.697333 Quilicura - 2012 - announced

2015 - launched

up to 20 employees expected. A million dollar investment plan to increase capacity at Quilicura was announced in 2018.
North America Reno (NV), USA 39°30′04″N 119°25′46″W / 39.5011558°N 119.4295537°W / 39.5011558; -119.4295537 Storey County - 2017 - 1,210 acres of land bought in the Tahoe Reno Industrial Center

2018 - announced

2018 November - project approved by the state of Nevada

North America Salt Lake City (UT), USA - Salt Lake City (us-west3) 2020 - launched
Asia Seoul, South Korea - Seoul

(asia-northeast3)

2020 - launched
Pacific Sydney, Australia - Sydney

(australia-southeast1)

2017 - launched
Middle East Tel Aviv, Israel - Tel Aviv (me-west1) 2022 - launched
North America The Dalles (OR), USA 45°37′57.04″N 121°12′8.16″W / 45.6325111°N 121.2022667°W / 45.6325111; -121.2022667 The Dalles Oregon (us-west1) 2006 - launched 80 full-time employees
North America Toronto, Canada - Toronto (northamerica-northeast2) 2021 - launched
Europe Turin, Italy 45°08′48″N 7°44′32″E / 45.146729°N 7.742147°E / 45.146729; 7.742147 - Turin (europe-west12) 2023 - launched
South America Vinhedo, São Paulo, Brazil São Paulo (southamerica-east1)
Europe Warsaw, Poland - Warsaw (europe-central2) 2019 - announced

2021 - launched

Asia Wenya, Jurong West, Singapore 1°21′04.8″N 103°42′35.2″E / 1.351333°N 103.709778°E / 1.351333; 103.709778 Singapore Singapore (asia-southeast1) 2011 - announced

2013 - launched

2015 - expanded

North America Widows Creek (Bridgeport) (AL), USA 34°54′48.4″N 85°44′53.1″W / 34.913444°N 85.748083°W / 34.913444; -85.748083 Jackson County - 2018 - broke ground
Europe Zürich, Switzerland 47°26′45″N 8°12′39″E / 47.4459257°N 8.2109085°E / 47.4459257; 8.2109085 - Zurich (europe-west6) 2018 - announced

2019 - launched

Europe Austria 2022 - announced
Europe Berlin, Germany Berlin (europe-west10) 2021 - announced 2023 August - launched
Middle East Dammam, Saudi Arabia 2021 - announced
Europe Athens, Greece 2022 - announced
North America Kansas City, Missouri 2019 - announced
Middle East Kuwait 2023 - announced
Asia Malaysia 2022 - announced
Pacific Auckland, New Zealand 2022 - announced
Europe Oslo, Norway 2022 - announced
North America Querétaro, Mexico 2022 - announced
Africa Johannesburg, South Africa Johannesburg (africa-south1) 2022 - announced2024 - launched
Europe Sweden 2022 - announced
Asia Tainan City, Taiwan - Taiwan

(asia-east1)

2019 September - announced
Asia Thailand 2022 - announced
Asia Yunlin County, Taiwan - Taiwan (asia-east1) 2020 September - announced
Europe Waltham Cross, Hertfordshire, UK 51°41′44″N 0°02′55″W / 51.695666°N 0.048728°W / 51.695666; -0.048728 2024 January - announced

Hardware

Original hardware

Google Data Centers 
Google's first production server rack, circa 1998

The original hardware (circa 1998) that was used by Google when it was located at Stanford University included:

  • Sun Microsystems Ultra II with dual 200 MHz processors, and 256 MB of RAM. This was the main machine for the original Backrub system.
  • 2 × 300 MHz dual Pentium II servers donated by Intel, they included 512 MB of RAM and 10 × 9 GB hard drives between the two. It was on these that the main search ran.
  • F50 IBM RS/6000 donated by IBM, included 4 processors, 512 MB of memory and 8 × 9 GB hard disk drives.
  • Two additional boxes included 3 × 9 GB hard drives and 6 x 4 GB hard disk drives respectively (the original storage for Backrub). These were attached to the Sun Ultra II.
  • SSD disk expansion box with another 8 × 9 GB hard disk drives donated by IBM.
  • Homemade disk box which contained 10 × 9 GB SCSI hard disk drives.

Production hardware

As of 2014, Google has used a heavily customized version of Debian Linux. They migrated from a Red Hat-based system incrementally in 2013.

The customization goal is to purchase CPU generations that offer the best performance per dollar, not absolute performance. How this is measured is unclear, but it is likely to incorporate running costs of the entire server, and CPU power consumption could be a significant factor. Servers as of 2009–2010 consisted of custom-made open-top systems containing two processors (each with several cores), a considerable amount of RAM spread over 8 DIMM slots housing double-height DIMMs, and at least two SATA hard disk drives connected through a non-standard ATX-sized power supply unit. The servers were open top so more servers could fit into a rack. According to CNET and a book by John Hennessy, each server had a novel 12-volt battery to reduce costs and improve power efficiency.

According to Google, their global data center operation electrical power ranges between 500 and 681 megawatts. The combined processing power of these servers might have reached from 20 to 100 petaflops in 2008.

Network topology

Details of the Google worldwide private networks are not publicly available, but Google publications make references to the "Atlas Top 10" report that ranks Google as the third largest ISP behind Level 3.

In order to run such a large network, with direct connections to as many ISPs as possible at the lowest possible cost, Google has a very open peering policy.

From this site, we can see that the Google network can be accessed from 67 public exchange points and 69 different locations across the world. As of May 2012, Google had 882 Gbit/s of public connectivity (not counting private peering agreements that Google has with the largest ISPs). This public network is used to distribute content to Google users as well as to crawl the internet to build its search indexes. The private side of the network is a secret, but a recent disclosure from Google indicate that they use custom built high-radix switch-routers (with a capacity of 128 × 10 Gigabit Ethernet port) for the wide area network. Running no less than two routers per datacenter (for redundancy) we can conclude that the Google network scales in the terabit per second range (with two fully loaded routers the bi-sectional bandwidth amount to 1,280 Gbit/s).

These custom switch-routers are connected to DWDM devices to interconnect data centers and point of presences (PoP) via dark fiber.

From a datacenter view, the network starts at the rack level, where 19-inch racks are custom-made and contain 40 to 80 servers (20 to 40 1U servers on either side, while new servers are 2U rackmount systems. Each rack has an Ethernet switch). Servers are connected via a 1 Gbit/s Ethernet link to the top of rack switch (TOR). TOR switches are then connected to a gigabit cluster switch using multiple gigabit or ten gigabit uplinks. The cluster switches themselves are interconnected and form the datacenter interconnect fabric (most likely using a dragonfly design rather than a classic butterfly or flattened butterfly layout).

From an operation standpoint, when a client computer attempts to connect to Google, several DNS servers resolve www.google.com into multiple IP addresses via Round Robin policy. Furthermore, this acts as the first level of load balancing and directs the client to different Google clusters. A Google cluster has thousands of servers, and once the client has connected to the server additional load balancing is done to send the queries to the least loaded web server. This makes Google one of the largest and most complex content delivery networks.

Google has numerous data centers scattered around the world. At least 12 significant Google data center installations are located in the United States. The largest known centers are located in The Dalles, Oregon; Atlanta, Georgia; Reston, Virginia; Lenoir, North Carolina; and Moncks Corner, South Carolina. In Europe, the largest known centers are in Eemshaven and Groningen in the Netherlands and Mons, Belgium. Google's Oceania Data Center is located in Sydney, Australia.

Data center network topology

To support fault tolerance, increase the scale of data centers and accommodate low-radix switches, Google has adopted various modified Clos topologies in the past.

Project 02

Google Data Centers 
Google data center in The Dalles, Oregon

One of the largest Google data centers is located in the town of The Dalles, Oregon, on the Columbia River, approximately 80 miles (129 km) from Portland. Codenamed "Project 02", the complex was built in 2006 and is approximately the size of two American football fields, with cooling towers four stories high. The site was chosen to take advantage of inexpensive hydroelectric power, and to tap into the region's large surplus of fiber optic cable, a remnant of the dot-com boom. A blueprint of the site appeared in 2008.

Summa papermill

In February 2009, Stora Enso announced that they had sold the Summa paper mill in Hamina, Finland to Google for 40 million Euros. Google invested 200 million euros on the site to build a data center and announced additional 150 million euro investment in 2012. Google chose this location due to the availability and proximity of renewable energy sources.

Floating data centers

In 2013, the press revealed the existence of Google's floating data centers along the coasts of the states of California (Treasure Island's Building 3) and Maine. The development project was maintained under tight secrecy. The data centers are 250 feet long, 72 feet wide, 16 feet deep. The patent for an in-ocean data center cooling technology was bought by Google in 2009 (along with a wave-powered ship-based data center patent in 2008). Shortly thereafter, Google declared that the two massive and secretly-built infrastructures were merely "interactive learning centers, [...] a space where people can learn about new technology."

Google halted work on the barges in late 2013 and began selling off the barges in 2014.

Software

Most of the software stack that Google uses on their servers was developed in-house. According to a well-known Google employee, C++, Java, Python and (more recently) Go are favored over other programming languages. For example, the back end of Gmail is written in Java and the back end of Google Search is written in C++. Google has acknowledged that Python has played an important role from the beginning, and that it continues to do so as the system grows and evolves.

The software that runs the Google infrastructure includes:

  • Google Web Server (GWS) – custom Linux-based Web server that Google uses for its online services.
  • Storage systems:
    • Google File System and its successor, Colossus
    • Bigtable – structured storage built upon GFS/Colossus
    • Spanner – planet-scale database, supporting externally-consistent distributed transactions
    • Google F1 – a distributed, quasi-SQL DBMS based on Spanner, substituting a custom version of MySQL.
  • Chubby lock service
  • MapReduce and Sawzall programming language
  • Indexing/search systems:
    • TeraGoogle – Google's large search index (launched in early 2006)
    • Caffeine (Percolator) – continuous indexing system (launched in 2010).
    • Hummingbird – major search index update, including complex search and voice search.
  • Borg declarative process scheduling software

Google has developed several abstractions which it uses for storing most of its data:

  • Protocol Buffers – "Google's lingua franca for data", a binary serialization format which is widely used within the company.
  • SSTable (Sorted Strings Table) – a persistent, ordered, immutable map from keys to values, where both keys and values are arbitrary byte strings. It is also used as one of the building blocks of Bigtable.
  • RecordIO – a sequence of variable sized records.

Software development practices

Most operations are read-only. When an update is required, queries are redirected to other servers, so as to simplify consistency issues. Queries are divided into sub-queries, where those sub-queries may be sent to different ducts in parallel, thus reducing the latency time.

To lessen the effects of unavoidable hardware failure, software is designed to be fault tolerant. Thus, when a system goes down, data is still available on other servers, which increases reliability.

Search infrastructure

Google Data Centers 
Google data center in The Dalles, Oregon

Index

Like most search engines, Google indexes documents by building a data structure known as inverted index. Such an index obtains a list of documents by a query word. The index is very large due to the number of documents stored in the servers.

The index is partitioned by document IDs into many pieces called shards. Each shard is replicated onto multiple servers. Initially, the index was being served from hard disk drives, as is done in traditional information retrieval (IR) systems. Google dealt with the increasing query volume by increasing number of replicas of each shard and thus increasing number of servers. Soon they found that they had enough servers to keep a copy of the whole index in main memory (although with low replication or no replication at all), and in early 2001 Google switched to an in-memory index system. This switch "radically changed many design parameters" of their search system, and allowed for a significant increase in throughput and a large decrease in latency of queries.

In June 2010, Google rolled out a next-generation indexing and serving system called "Caffeine" which can continuously crawl and update the search index. Previously, Google updated its search index in batches using a series of MapReduce jobs. The index was separated into several layers, some of which were updated faster than the others, and the main layer wouldn't be updated for as long as two weeks. With Caffeine, the entire index is updated incrementally on a continuous basis. Later Google revealed a distributed data processing system called "Percolator" which is said to be the basis of Caffeine indexing system.

Server types

Google's server infrastructure is divided into several types, each assigned to a different purpose:

  • Web servers coordinate the execution of queries sent by users, then format the result into an HTML page. The execution consists of sending queries to index servers, merging the results, computing their rank, retrieving a summary for each hit (using the document server), asking for suggestions from the spelling servers, and finally getting a list of advertisements from the ad server.
  • Data-gathering servers are permanently dedicated to spidering the Web. Google's web crawler is known as GoogleBot. They update the index and document databases and apply Google's algorithms to assign ranks to pages.
  • Each index server contains a set of index shards. They return a list of document IDs ("docid"), such that documents corresponding to a certain docid contain the query word. These servers need less disk space, but suffer the greatest CPU workload.
  • Document servers store documents. Each document is stored on dozens of document servers. When performing a search, a document server returns a summary for the document based on query words. They can also fetch the complete document when asked. These servers need more disk space.
  • Ad servers manage advertisements offered by services like AdWords and AdSense.
  • Spelling servers make suggestions about the spelling of queries.

Security

External videos
Google Data Centers 
YouTube
Google Data Centers  Google Data Center Security: 6 Layers Deep

In October 2013, The Washington Post reported that the U.S. National Security Agency intercepted communications between Google's data centers, as part of a program named MUSCULAR. This wiretapping was made possible because, at the time, Google did not encrypt data passed inside its own network. This was rectified when Google began encrypting data sent between data centers in 2013.

Environmental impact

Google Data Centers 
Google data center in Mayes County, Oklahoma at MidAmerica Industrial Park

Google's most efficient data center runs at 35 °C (95 °F) using only fresh air cooling, requiring no electrically powered air conditioning.

In December 2016, Google announced that—starting in 2017—it would purchase enough renewable energy to match 100% of the energy usage of its data centers and offices. The commitment will make Google "the world's largest corporate buyer of renewable power, with commitments reaching 2.6 gigawatts (2,600 megawatts) of wind and solar energy".

References

Tags:

Google Data Centers LocationsGoogle Data Centers HardwareGoogle Data Centers SoftwareGoogle Data Centers Search infrastructureGoogle Data Centers SecurityGoogle Data Centers Environmental impactGoogle Data Centers Further readingGoogle Data Centers

🔥 Trending searches on Wiki English:

Rebel WilsonErling HaalandLarry NassarEFL ChampionshipEmily BluntAndrew Scott (actor)Los AngelesThe UndertakerRobert DurstBlue Lights (2023 TV series)List of American films of 2024Bob GrahamList of United States cities by populationBacklash FrancePaul McCartneyFreddie MercuryXXXTentacionInna Lillahi wa inna ilayhi raji'unBlack holeJessica GunningAl Hilal SFCRoyce FreemanSelena GomezRishi SunakCatBangladesh2023–24 UEFA Champions LeagueEstadi Olímpic Lluís Companys2024 Iranian strikes in IsraelEl Clásico2024 in filmScottie SchefflerCanvaUnited KingdomMarjorie Taylor GreeneWish (film)Jeremie FrimpongHeartbreak High (2022 TV series)Kobe BryantRhea Ripley2022–23 UEFA Champions LeagueKim KardashianMamitha BaijuKalanithi MaranSouth AfricaAli KhameneiJai Shri RamUnited Arab Emirates2024 Indian general election in AssamLindsey BuckinghamAnyone but YouIsidor StrausDune MessiahJose Alvarado (basketball)Dev PatelLucas VázquezRahul GandhiRussian invasion of UkraineKorean WarJonathan BrandisApril 17Sage StalloneRobin WilliamsPhilippinesNBA playoffsImaginary (film)Dwayne JohnsonKatherine MaherWomen's National Basketball AssociationKeegan MurrayKakáJake GyllenhaalList of country calling codesDerek UnderwoodNicola PeltzTimothée ChalametAnna (2019 feature film)UEFA Champions League🡆 More