[HTML payload içeriği buraya]
32.7 C
Jakarta
Sunday, May 17, 2026

How A.I. Is Altering the Means the World Builds Computer systems


That is essentially the most basic change to computing for the reason that early days of the World Huge Net. Simply as corporations fully rebuilt their laptop methods to accommodate the brand new industrial web within the Nineties, they’re now rebuilding from the underside up — from tiny elements to the best way that computer systems are housed and powered — to accommodate synthetic intelligence.

Massive tech corporations have constructed laptop knowledge facilities everywhere in the world for 20 years. The facilities have been full of computer systems to deal with the web site visitors flooding into the businesses’ web providers, together with search engines like google, electronic mail purposes and e-commerce websites.

However these services had been lightweights in contrast with what’s coming. Again in 2006, Google opened its first knowledge heart in The Dalles, Ore., spending an estimated $600 million to finish the power. In January, OpenAI and several other companions introduced a plan to spend roughly $100 billion on new knowledge facilities, starting with a campus in Texas. They plan to finally pump a further $400 billion into this and different services throughout america.

The change in computing is reshaping not simply expertise but in addition finance, vitality and communities. Non-public fairness companies are plowing cash into knowledge heart corporations. Electricians are flocking to areas the place the services are being erected. And in some locations, locals are pushing again towards the tasks, fearful that they may convey extra hurt than good.

For now, tech corporations are asking for extra computing energy and extra electrical energy than the world can present. OpenAI hopes to lift tons of of billions of {dollars} to assemble laptop chip factories within the Center East. Google and Amazon not too long ago struck offers to construct and deploy a brand new era of nuclear reactors. They usually need to do it quick.

Google’s A.I. chips on a circuit board. The corporate wants hundreds of those chips to construct its chatbots and different A.I. applied sciences.

Christie Hemm Klok for The New York Instances

The larger-is-better mantra was challenged in December when a tiny Chinese language firm, DeepSeek, stated it had constructed one of many world’s strongest A.I. methods utilizing far fewer laptop chips than many consultants thought doable. That raised questions on Silicon Valley’s frantic spending.

U.S. tech giants had been unfazed. The wildly formidable objective of many of those corporations is to create synthetic basic intelligence, or A.G.I. — a machine that may do something the human mind can do — and so they nonetheless consider that having extra computing energy is crucial to get there.

Amazon, Meta, Microsoft, and Google’s guardian firm, Alphabet, not too long ago indicated that their capital spending — which is primarily used to construct knowledge facilities — may prime a mixed $320 billion this 12 months. That’s greater than twice what they spent two years in the past.

The New York Instances visited 5 new knowledge heart campuses in California, Utah, Texas and Oklahoma and spoke with greater than 50 executives, engineers, entrepreneurs and electricians to inform the story of the tech business’s insatiable starvation for this new sort of computing.

“What was in all probability going to occur over the following decade has been compressed right into a interval of simply two years,” Sundar Pichai, Google’s chief government, stated in an interview with The Instances. “A.I. is the accelerant.”

New laptop chips for brand new A.I.

The enormous leap ahead in computing for A.I. was pushed by a tiny ingredient: the specialised laptop chips referred to as graphics processing items, or GPUs.

Firms just like the Silicon Valley chipmaker Nvidia initially designed these chips to render graphics for video video games. However GPUs had a knack for operating the maths that powers what are referred to as neural networks, which might study abilities by analyzing giant quantities of information. Neural networks are the idea of chatbots and different main A.I. applied sciences.

How A.I. Fashions Are Skilled

By analyzing huge datasets, algorithms can study to differentiate between photos, in what’s referred to as machine studying. The instance beneath demonstrates the coaching technique of an A.I. mannequin to determine a picture of a flower based mostly on current flower photos.

Sources: IBM and Cloudflare

The New York Instances

Prior to now, computing largely relied on chips referred to as central processing items, or CPUs. These may do many issues, together with the easy math that powers neural networks.

However GPUs can do that math quicker — quite a bit quicker. At any given second, a standard chip can do a single calculation. In that very same second, a GPU can do hundreds. Laptop scientists name this parallel processing. And it means neural networks can analyze extra knowledge.

“These are very totally different from chips used to simply serve up an online web page,” stated Vipul Ved Prakash, the chief government of Collectively AI, a tech consultancy. “They run hundreds of thousands of calculations as a method for machines to ‘assume’ about an issue.”

So tech corporations began utilizing more and more giant numbers of GPUs to construct more and more highly effective A.I. applied sciences.

Distinction between CPU and GPU-powered computer systems

Sources: Nvidia, IBM and Cloudflare

The New York Instances

Alongside the best way, Nvidia rebuilt its GPUs particularly for A.I., packing extra transistors into every chip to run much more calculations with every passing second. In 2013, Google started constructing its personal A.I. chips.

These Google and Nvidia chips weren’t designed to run laptop working methods and couldn’t deal with the varied features for working a Home windows laptop computer or an iPhone. However working collectively, they accelerated the creation of A.I.

“The outdated mannequin lasted for about 50 years,” stated Norm Jouppi, a Google engineer who oversees the corporate’s effort to construct new silicon chips for A.I. “Now, we have now a very totally different method of doing issues.”

The nearer the chips, the higher.

It isn’t simply the chips which can be totally different. To get essentially the most out of GPUs, tech corporations should pace the circulate of digital knowledge among the many chips.

“Each GPU wants to speak to each different GPU as quick as doable,” stated Dave Driggers, the chief expertise officer at Cirrascale Cloud Providers, which operates an information heart in Austin, Texas, for the Allen Institute for Synthetic Intelligence, a distinguished A.I. analysis lab.

The nearer the chips are to at least one one other, the quicker they’ll work. So corporations are packing as many chips right into a single knowledge heart as they’ll. They’ve additionally developed new {hardware} and cabling to quickly stream knowledge from chip to chip.

Meta’s Eagle Mountain knowledge heart sits in a valley beneath Utah’s Lake Mountains, south of Salt Lake Metropolis. Meta broke floor on this constructing after the A.I. growth erupted.

Christie Hemm Klok for The New York Instances

That’s altering how knowledge facilities — that are primarily massive buildings crammed with racks of computer systems stacked on prime of each other — work.

In 2021, earlier than the A.I. growth, Meta opened two knowledge facilities an hour south of Salt Lake Metropolis and was constructing three extra there. These services — every the dimensions of the Empire State Constructing, laid on its aspect throughout the desert — would assist energy the corporate’s social media apps, equivalent to Fb and Instagram.

However after OpenAI launched ChatGPT in 2022, Meta re-evaluated its A.I. plans. It needed to cram hundreds of GPUs into a brand new knowledge heart so they might churn via weeks and even months of calculations wanted to construct a single neural community and advance the corporate’s A.I.

“Every thing should perform as one big, data-center-sized supercomputer,” stated Rachel Peterson, Meta’s vp of information facilities. “That may be a complete totally different equation.”

Inside months, Meta broke floor on a sixth and seventh Utah knowledge heart beside the opposite 5. In these 700,000-square-foot services, technicians stuffed every rack with {hardware} used to coach A.I., sliding in boxy machines full of GPUs that may price tens of hundreds of {dollars}.

In 2023, Meta incurred a $4.2 billion restructuring cost, partly to revamp lots of its future knowledge heart tasks for A.I. Its exercise was emblematic of a change occurring throughout the tech business.

A.I. machines want extra electrical energy. Way more.

New knowledge facilities full of GPUs meant new electrical energy calls for — a lot in order that the urge for food for energy would undergo the roof.

In December 2023, Cirrascale leased a 139,000-square-foot conventional knowledge heart in Austin that drew on 5 megawatts of electrical energy, sufficient to energy about 3,600 common American houses. Inside, computer systems had been organized in about 80 rows. Then the corporate ripped out the outdated computer systems to transform the power for A.I.

The 5 megawatts that used to energy a constructing stuffed with CPUs is now sufficient to run simply eight to 10 rows of computer systems full of GPUs. Cirrascale can increase to about 50 megawatts of electrical energy from the grid, however even that might not fill the information heart with GPUs.

And that’s nonetheless on the small aspect. OpenAI goals to construct about 5 knowledge facilities that prime {the electrical} use of about three million households.

Cirrascale’s knowledge heart in Austin, Texas, attracts on 5 megawatts of electrical energy, which might energy eight to 10 rows of computer systems full of GPUs.

Christie Hemm Klok for The New York Instances

It’s not simply that these knowledge facilities have extra gear packed right into a tighter house. The pc chips that A.I. revolves round want way more electrical energy than conventional chips. A typical CPU wants about 250 to 500 watts to run, whereas GPUs use as much as 1,000 watts.

Constructing an information heart is finally a negotiation with the native utility. How a lot energy can it present? At what price? If it should increase {the electrical} grid with hundreds of thousands of {dollars} in new gear, who pays for the upgrades?

Information facilities consumed about 4.4 p.c of whole electrical energy in america in 2023, or greater than twice as a lot energy because the services used to mine cryptocurrencies. That might triple by 2028, based on a December report printed by the Division of Vitality.

Energy consumption by A.I. knowledge facilities

The Vitality Division estimates that A.I. servers in knowledge facilities may eat as a lot as 326 terawatt-hours by 2028, practically eight instances what they utilized in 2023.

Supply: Lawrence Berkeley Nationwide Laboratory, Vitality Division

The New York Instances

“Time is the forex within the business proper now,” stated Arman Shehabi, a researcher on the Lawrence Berkeley Nationwide Laboratory who led the report. There’s a rush to maintain constructing, he stated, and “I don’t see this slowing down within the subsequent few years.”

Information heart operators are actually having bother discovering electrical energy in america. In areas like Northern Virginia — the world’s largest hub of information facilities due to its proximity to underwater cables that shuttle knowledge to and from Europe — these corporations have all however exhausted the obtainable electrical energy.

Some A.I. giants are turning to nuclear energy. Microsoft is restarting the Three Mile Island nuclear plant in Pennsylvania.

Others are taking totally different routes. Elon Musk and xAI, his A.I. start-up, not too long ago bypassed clear vitality in favor of a faster resolution: putting in their very own fuel generators at a brand new knowledge heart in Memphis.

“My conversations have gone from ‘The place can we get some state-of-the-art chips?’ to ‘The place can we get some electrical energy?’” stated David Katz, a associate with Radical Ventures, a enterprise capital agency that invests in A.I.

A.I. will get so sizzling, solely water can cool it down.

These unusually dense A.I. methods have led to a different change: a special method of cooling computer systems.

A.I. methods can get highly regarded. As air circulates from the entrance of a rack and crosses the chips crunching calculations, it heats up. At Cirrascale’s Austin knowledge heart, the temperature round one rack began at 71.2 levels Fahrenheit on the entrance and ended up at 96.9 levels on the again aspect.

If a rack isn’t correctly cooled down, the machines — and doubtlessly the entire knowledge heart — are susceptible to catching hearth.

Simply exterior Pryor, a farm-and-cattle city within the northeast nook of Oklahoma, Google is fixing this downside on a large scale.

13 Google knowledge facilities stand up from the grassy flatlands. This campus holds tens of hundreds of racks of machines and makes use of tons of of megawatts of electrical energy streaming from metal-and-wire energy stations put in between the concrete buildings. To maintain the machines from overheating, Google pumps chilly water via all 13 buildings.

Prior to now, Google’s water pipes ran via empty aisles beside the racks of computer systems. Because the chilly water moved via the pipes, it absorbed the warmth from the encompassing air. However when the racks are full of A.I. chips, the water isn’t shut sufficient to soak up the additional warmth.

Supply: SimScale thermodynamics

The New York Instances

Google now runs its water pipes proper up subsequent to the chips. Solely then can the water take up the warmth and hold the chips working.

Supply: SimScale thermodynamics

The New York Instances

Pumping water via an information heart crammed with electrical gear will be dangerous since water can leak from the pipes onto the pc {hardware}. So Google treats its water with chemical substances that make it much less more likely to conduct electrical energy — and fewer more likely to harm the chips.

As soon as the water absorbs the warmth from all these chips, tech corporations should additionally discover methods of cooling the water again down.

In lots of instances, they do that utilizing big towers sitting on the roof of the information heart. A number of the water evaporates from these towers, which cools the remainder of it, a lot as persons are cooled once they sweat and the sweat evaporates from their pores and skin.

“That’s what we name free cooling — the evaporation that occurs naturally on a cool, dry morning,” stated Joe Kava, Google’s vp of information facilities.

Inside a Google knowledge heart, which is full of computer systems that use Google’s A.I. chips.

Christie Hemm Klok for The New York Instances

Google and different corporations that use this method should hold replenishing the water that pumps via the information heart, which might pressure native water provides.

Google knowledge facilities consumed 6.1 billion gallons of water in 2023, up 17 p.c from the earlier 12 months. In California, a state that faces drought, greater than 250 knowledge facilities eat billions of gallons of water yearly, elevating alarm bells amongst native officers.

Some corporations, together with Cirrascale, use huge chillers — primarily air-conditioners — to chill their water as a substitute. That reduces strain on the native water provide, as a result of they reuse just about all the water. However the course of requires extra electrical energy.

There may be little finish in sight. Final 12 months, Google broke floor on 11 knowledge facilities in South Carolina, Indiana, Missouri and elsewhere. Meta stated its latest facility, in Richland Parish, La., could be sufficiently big to cowl most of Central Park, Midtown Manhattan, Greenwich Village and the Decrease East Aspect.

“This might be a defining 12 months for AI,” Mark Zuckerberg, Meta’s chief government, stated in January in a Fb submit that concluded, “Let’s go construct!”

Related Articles

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Latest Articles