
<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>AI infrastructure growth &#8211; The Milli Chronicle</title>
	<atom:link href="https://www.millichronicle.com/tag/ai-infrastructure-growth/feed" rel="self" type="application/rss+xml" />
	<link>https://www.millichronicle.com</link>
	<description>Factual Version of a Story</description>
	<lastBuildDate>Sat, 24 Jan 2026 20:12:21 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	

<image>
	<url>https://media.millichronicle.com/2018/11/12122950/logo-m-01-150x150.png</url>
	<title>AI infrastructure growth &#8211; The Milli Chronicle</title>
	<link>https://www.millichronicle.com</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>Intel’s Long-Term AI Opportunity Remains Intact as Supply Constraints Highlight Demand Strength</title>
		<link>https://www.millichronicle.com/2026/01/62466.html</link>
		
		<dc:creator><![CDATA[NewsDesk Milli Chronicle]]></dc:creator>
		<pubDate>Sat, 24 Jan 2026 20:12:21 +0000</pubDate>
				<category><![CDATA[Featured]]></category>
		<category><![CDATA[News]]></category>
		<category><![CDATA[World]]></category>
		<category><![CDATA[advanced chip manufacturing]]></category>
		<category><![CDATA[AI data centers]]></category>
		<category><![CDATA[AI hardware demand]]></category>
		<category><![CDATA[AI infrastructure growth]]></category>
		<category><![CDATA[global semiconductor industry]]></category>
		<category><![CDATA[Intel AI chips demand]]></category>
		<category><![CDATA[Intel CEO Lip-Bu Tan]]></category>
		<category><![CDATA[Intel data center processors]]></category>
		<category><![CDATA[Intel foundry plans]]></category>
		<category><![CDATA[Intel manufacturing roadmap]]></category>
		<category><![CDATA[Intel stock outlook]]></category>
		<category><![CDATA[Intel technology innovation]]></category>
		<category><![CDATA[Intel turnaround strategy]]></category>
		<category><![CDATA[long-term tech investing]]></category>
		<category><![CDATA[market volatility stocks]]></category>
		<category><![CDATA[PC chip recovery]]></category>
		<category><![CDATA[semiconductor investment trends]]></category>
		<category><![CDATA[semiconductor supply constraints]]></category>
		<category><![CDATA[server chip market]]></category>
		<guid isPermaLink="false">https://millichronicle.com/?p=62466</guid>

					<description><![CDATA[Intel’s recent share dip reflects short-term supply challenges rather than weakening fundamentals, underscoring strong demand for its data-center chips as]]></description>
										<content:encoded><![CDATA[
<blockquote class="wp-block-quote">
<p>Intel’s recent share dip reflects short-term supply challenges rather than weakening fundamentals, underscoring strong demand for its data-center chips as the company advances its broader turnaround strategy.</p>
</blockquote>



<p>Intel’s latest stock movement has drawn attention, but the underlying story remains one of rising demand and structural change rather than decline.</p>



<p>The recent pullback highlights how strong interest in AI-linked data-center chips has temporarily outpaced supply, a sign of momentum rather than market rejection.</p>



<p>After spending years on the sidelines of the artificial intelligence boom, Intel is now experiencing a meaningful surge in demand for its traditional server processors.</p>



<p>These chips play a critical supporting role alongside advanced graphics processors in modern data centers, anchoring Intel firmly in the AI ecosystem.</p>



<p>Investor enthusiasm around Intel’s comeback has been building steadily over the past year.</p>



<p>Major backing from the U.S. government, global technology investors, and strategic partners has reinforced confidence in the company’s long-term vision.</p>



<p>Intel’s shares delivered exceptional gains over the past year, outperforming many peers in the semiconductor space. The recent volatility follows an extended rally, making some consolidation a natural part of the market cycle.</p>



<p>Supply constraints, while challenging in the near term, signal how sharply demand has accelerated. Intel’s factories are operating at high utilization levels, reflecting strong customer interest across enterprise and cloud markets.</p>



<p>Company leadership has been transparent about these near-term pressures. Executives have indicated that supply availability is expected to improve as early as the second quarter, easing bottlenecks and supporting delivery timelines.</p>



<p>Industry analysts broadly agree that the tightest part of the supply cycle is likely temporary. Several forecasts suggest capacity constraints should bottom out by early spring, setting the stage for smoother operations later in the year.</p>



<p>Intel’s role in data centers remains strategically important as AI workloads expand globally. Even as specialized processors gain attention, server CPUs remain essential for managing, coordinating, and scaling AI systems.</p>



<p>Beyond data centers, Intel continues to position itself for a recovery in the personal computer market. Its upcoming PC chip platforms are designed to reignite consumer and enterprise upgrades after a prolonged slowdown.</p>



<p>Memory market dynamics have added another layer of complexity to near-term forecasts. However, these industry-wide pressures are expected to normalize, benefiting large, diversified players with scale and pricing power.</p>



<p>Under CEO Lip-Bu Tan, Intel’s turnaround strategy emphasizes focus, efficiency, and disciplined investment. Cost controls and a refined manufacturing roadmap are intended to strengthen margins and execution over time.</p>



<p>The company has also taken a more measured approach to contract manufacturing ambitions. This recalibration allows Intel to prioritize internal innovation while selectively engaging external customers.</p>



<p>Investor attention remains high around Intel’s advanced manufacturing technologies. Ongoing evaluations of next-generation process nodes suggest growing industry interest in Intel’s technical capabilities.</p>



<p>While some expectations around immediate customer commitments may have been optimistic, the evaluation phase itself reflects credibility. Such assessments often precede deeper partnerships once production readiness improves.</p>



<p>Market reactions to quarterly guidance often reflect short-term sentiment rather than long-term value. Intel’s leadership continues to emphasize progress over quarters and years, not weeks.</p>



<p>The broader semiconductor landscape remains highly competitive, but Intel’s scale offers resilience. Few companies combine design expertise, manufacturing depth, and ecosystem reach at Intel’s level.</p>



<p>Global demand for computing power continues to rise, driven by AI, cloud services, and digital transformation. Intel’s product portfolio positions it to participate across multiple growth vectors rather than a single niche.</p>



<p>Short-term stock volatility is common during major corporate transformations. History shows that companies executing complex turnarounds often face uneven market reactions before stability returns.</p>



<p>Intel’s renewed momentum, supported by policy backing and strategic investment, remains a key differentiator. As supply constraints ease, investors may refocus on demand strength and execution progress.</p>



<p>Overall, the current phase represents adjustment rather than setback. Intel’s long-term opportunity in AI-driven infrastructure and computing remains firmly in place.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Nvidia explores expanding H200 chip production to meet growing China demand</title>
		<link>https://www.millichronicle.com/2025/12/60702.html</link>
		
		<dc:creator><![CDATA[NewsDesk Milli Chronicle]]></dc:creator>
		<pubDate>Sat, 13 Dec 2025 19:23:31 +0000</pubDate>
				<category><![CDATA[Featured]]></category>
		<category><![CDATA[News]]></category>
		<category><![CDATA[World]]></category>
		<category><![CDATA[advanced semiconductor manufacturing]]></category>
		<category><![CDATA[AI chip production]]></category>
		<category><![CDATA[AI hardware expansion]]></category>
		<category><![CDATA[AI infrastructure growth]]></category>
		<category><![CDATA[AI innovation China]]></category>
		<category><![CDATA[AI technology leadership]]></category>
		<category><![CDATA[Alibaba AI orders]]></category>
		<category><![CDATA[ByteDance H200]]></category>
		<category><![CDATA[Chinese AI market]]></category>
		<category><![CDATA[cloud computing AI]]></category>
		<category><![CDATA[enterprise AI solutions]]></category>
		<category><![CDATA[global AI technology]]></category>
		<category><![CDATA[high-performance AI chips]]></category>
		<category><![CDATA[Hopper AI chips]]></category>
		<category><![CDATA[international chip exports]]></category>
		<category><![CDATA[Nvidia China demand]]></category>
		<category><![CDATA[Nvidia chip supply]]></category>
		<category><![CDATA[Nvidia H200]]></category>
		<category><![CDATA[Rubin chip transition]]></category>
		<category><![CDATA[TSMC 4nm process]]></category>
		<guid isPermaLink="false">https://millichronicle.com/?p=60702</guid>

					<description><![CDATA[Nvidia moves to scale up H200 AI chip output as Chinese interest surges, highlighting robust global demand and strategic supply]]></description>
										<content:encoded><![CDATA[
<blockquote class="wp-block-quote">
<p>Nvidia moves to scale up H200 AI chip output as Chinese interest surges, highlighting robust global demand and strategic supply management for advanced AI technologies.</p>
</blockquote>



<p>Nvidia is evaluating an increase in production capacity for its high-performance H200 AI chips after Chinese orders exceeded current supply expectations.</p>



<p>The U.S. government recently approved the export of H200 chips to China under a 25% fee, enabling Nvidia to serve authorized Chinese clients while maintaining commitments to U.S. customers.</p>



<p>Chinese technology leaders, including Alibaba and ByteDance, have expressed strong interest in large H200 orders, reflecting the chip’s leading-edge performance in AI applications.</p>



<p>While demand is robust, final approval from the Chinese government is still pending, and discussions continue regarding potential regulatory conditions for imports.</p>



<p>The H200, manufactured using TSMC’s advanced 4nm process, represents the fastest offering from Nvidia’s Hopper generation, providing unmatched computational power for AI workloads.</p>



<p>Its capabilities are approximately six times stronger than Nvidia’s previous H20 chip tailored for the Chinese market, making it a highly sought-after resource for AI innovation.</p>



<p>Nvidia has reassured clients that expanding supply to China will not disrupt deliveries to U.S. customers, demonstrating careful management of global production and logistics.</p>



<p>The company is also transitioning to its next-generation Rubin chips while balancing ongoing H200 production to meet international demand and maintain strategic market leadership.</p>



<p>Emergency discussions within China have included proposals to link H200 imports with domestic chip purchases, aiming to support local AI industry growth alongside international technology adoption.</p>



<p>Nvidia’s proactive engagement with Chinese clients highlights its responsiveness to market demand and commitment to supporting enterprise-level AI deployments.</p>



<p>Investors and industry observers view the potential production expansion positively, as it underscores Nvidia’s role in supplying cutting-edge AI infrastructure to leading technology companies worldwide.</p>



<p>The H200’s deployment enhances cloud computing, data analytics, and AI research capabilities, allowing enterprises to accelerate innovation and improve efficiency in complex computational tasks.</p>



<p>By scaling capacity, Nvidia positions itself to meet unprecedented demand in the AI sector, reinforcing its status as a global leader in high-performance computing solutions.</p>



<p>The company’s strategic planning ensures that new production lines integrate seamlessly with existing manufacturing schedules while maintaining quality and reliability standards.</p>



<p>Nvidia’s engagement with TSMC and global partners demonstrates collaboration at the highest levels of semiconductor manufacturing to meet surging international orders.</p>



<p>Chinese interest in the H200 underscores the region’s commitment to adopting world-class AI technology while fostering domestic innovation and competitive capabilities.</p>



<p>Expanding H200 availability can help accelerate AI-driven research, enterprise deployment, and technological advancement across industries such as finance, healthcare, and e-commerce.</p>



<p>Nvidia’s careful navigation of regulatory approvals and supply chain logistics illustrates its expertise in balancing global demand with strategic growth objectives in the AI market.</p>



<p>The company’s continued innovation in AI chips, combined with measured capacity expansion, strengthens its competitive positioning and long-term growth prospects.</p>



<p>With rising international interest and carefully managed production plans, Nvidia is poised to deliver transformative AI capabilities to clients while driving the next wave of global AI development.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Micron to Wind Down ‘Crucial’ Consumer Memory Line as Company Refocuses on High-Growth Segments</title>
		<link>https://www.millichronicle.com/2025/12/60205.html</link>
		
		<dc:creator><![CDATA[NewsDesk Milli Chronicle]]></dc:creator>
		<pubDate>Wed, 03 Dec 2025 17:55:33 +0000</pubDate>
				<category><![CDATA[Featured]]></category>
		<category><![CDATA[News]]></category>
		<category><![CDATA[World]]></category>
		<category><![CDATA[AI data-center memory demand]]></category>
		<category><![CDATA[AI infrastructure growth]]></category>
		<category><![CDATA[cloud computing hardware trends]]></category>
		<category><![CDATA[Crucial memory brand shutdown]]></category>
		<category><![CDATA[enterprise memory strategy]]></category>
		<category><![CDATA[global chip industry shift]]></category>
		<category><![CDATA[high-performance DRAM market]]></category>
		<category><![CDATA[memory supply reallocation]]></category>
		<category><![CDATA[Micron consumer exit]]></category>
		<category><![CDATA[SSD phaseout Micron]]></category>
		<guid isPermaLink="false">https://millichronicle.com/?p=60205</guid>

					<description><![CDATA[Micron is stepping away from its well-known Crucial consumer brand to prioritize soaring demand from data-center and AI clients, marking]]></description>
										<content:encoded><![CDATA[
<blockquote class="wp-block-quote">
<p> Micron is stepping away from its well-known Crucial consumer brand to prioritize soaring demand from data-center and AI clients, marking a major strategic shift in its global product lineup.</p>
</blockquote>



<p>Micron Technology announced that it will exit its long-running Crucial consumer memory business, a move that signals a significant reorientation toward high-growth enterprise and data-center markets driven by accelerating demand for advanced memory and storage.</p>



<p>The company confirmed that the phaseout will include discontinuing the sale of all consumer-branded products through global retailers, online marketplaces, and distribution channels connected to the Crucial brand.</p>



<p>Micron stated that shipments for existing consumer products will continue into early 2026, ensuring a transition period for partners, resellers, and consumers who rely on the long-established memory brand.</p>



<p>The company’s shares slipped slightly following the announcement, reflecting market reaction to the departure from a unit that has been part of Micron’s broader commercial footprint for years.</p>



<p>Executives emphasized that the shift comes at a critical moment for the industry, with artificial intelligence workloads fueling unprecedented demand for high-performance memory in data centers worldwide.</p>



<p>According to company leadership, rapidly expanding AI infrastructure requirements have placed pressure on supply allocation, prompting a renewed focus on segments with stronger growth trajectories and deeper strategic value.</p>



<p>Micron’s chief business officer highlighted that the company made what he called a difficult but necessary decision to exit the consumer line to better support global enterprise clients and technology partners currently driving the surge in next-generation memory development.</p>



<p>Industry analysts note that the Crucial brand—widely recognized for SSDs, DRAM modules, and portable storage—has played a major role in the consumer PC and DIY upgrade market, serving hobbyists and general users for more than two decades.</p>



<p>The discontinuation marks a transition away from lower-margin consumer categories in favor of opportunities linked to hyperscale computing, cloud platforms, AI training systems, and advanced data-processing environments.</p>



<p>Market observers say that as AI models grow more complex, companies like Micron face increasing pressure to allocate resources to memory technologies that can support the scaling needs of major global cloud providers.</p>



<p>The company’s decision comes at a time when enterprise customers are rapidly expanding investment in accelerated computing, putting further strain on supply chains for high-bandwidth memory products.</p>



<p>With more organizations now adopting AI-driven workloads, demand for top-tier DRAM and cutting-edge NAND solutions is expected to remain strong through 2026, reinforcing Micron’s strategic pivot.</p>



<p>Despite stepping away from the consumer market, Micron reaffirmed its commitment to supporting partners and ensuring continuity throughout the transition window, especially for regions where Crucial products remain widely used.</p>



<p>Industry experts suggest that while PC enthusiasts may view the move as the end of a familiar brand era, the broader strategy aligns with ongoing industry consolidation toward enterprise and AI-focused infrastructure.</p>



<p>The realignment is expected to enhance Micron’s ability to supply memory for servers, accelerators, and high-capacity storage systems that are increasingly central to global technological advancement.</p>



<p>As enterprise clients shift toward AI-optimized infrastructure, memory manufacturers are being compelled to prioritize segments that can support the pace of innovation and volume requirements associated with advanced computing.</p>



<p>The company’s pivot reflects a broader industry trend of repositioning product portfolios to capture long-term growth tied to the rapid evolution of artificial intelligence, cloud ecosystems, and high-performance processing.</p>



<p>Micron’s exit from the consumer space marks a milestone in its corporate strategy, reinforcing its focus on foundational technologies that underpin the next wave of digital transformation.</p>



<p>The announcement also underscores the growing divide between consumer-grade offerings and enterprise-level solutions, particularly as memory requirements accelerate across global AI platforms.</p>



<p>While the Crucial brand winds down, Micron is positioning itself more aggressively within high-value markets where demand, innovation, and profitability are expected to expand significantly over the next several years.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Nvidia CFO Says $100 Billion OpenAI Investment Plan Still Not Finalized</title>
		<link>https://www.millichronicle.com/2025/12/60155.html</link>
		
		<dc:creator><![CDATA[NewsDesk Milli Chronicle]]></dc:creator>
		<pubDate>Tue, 02 Dec 2025 20:32:58 +0000</pubDate>
				<category><![CDATA[Featured]]></category>
		<category><![CDATA[News]]></category>
		<category><![CDATA[World]]></category>
		<category><![CDATA[AI ecosystem deals]]></category>
		<category><![CDATA[AI infrastructure growth]]></category>
		<category><![CDATA[AI investment news]]></category>
		<category><![CDATA[Anthropic investment]]></category>
		<category><![CDATA[cloud computing capacity]]></category>
		<category><![CDATA[generative AI expansion]]></category>
		<category><![CDATA[GPU demand]]></category>
		<category><![CDATA[high-performance chips]]></category>
		<category><![CDATA[Nvidia CFO comments]]></category>
		<category><![CDATA[Nvidia chip bookings]]></category>
		<category><![CDATA[Nvidia OpenAI deal]]></category>
		<category><![CDATA[OpenAI partnership]]></category>
		<category><![CDATA[technology conference updates]]></category>
		<guid isPermaLink="false">https://millichronicle.com/?p=60155</guid>

					<description><![CDATA[Nvidia signals that its headline-making multibillion-dollar investment proposal with OpenAI remains under negotiation, as the chipmaker navigates growing scrutiny over]]></description>
										<content:encoded><![CDATA[
<blockquote class="wp-block-quote">
<p>Nvidia signals that its headline-making multibillion-dollar investment proposal with OpenAI remains under negotiation, as the chipmaker navigates growing scrutiny over large AI-ecosystem partnerships and expanding demand for advanced computing power.</p>
</blockquote>



<p>Nvidia has clarified that its proposed investment of up to $100 billion into OpenAI is still not finalized, despite widespread industry attention on the potential scale and implications of the arrangement.</p>



<p>The company’s chief financial officer, Colette Kress, addressed the topic at a major technology and AI conference, saying discussions with the AI startup are ongoing and no definitive agreement has yet been completed.</p>



<p>Kress’ remarks come at a time when the relationship between major chipmakers and leading AI developers is increasingly under the spotlight, particularly as companies form deep, interdependent partnerships.</p>



<p>The proposed deal between Nvidia and OpenAI has drawn significant attention given the size of the potential investment and the growing influence of both companies in the global artificial intelligence landscape.</p>



<p>The initial framework outlined earlier this year involved a letter of intent signaling Nvidia&#8217;s readiness to deploy at least 10 gigawatts of computing capacity for OpenAI’s future infrastructure.</p>



<p>This scale of deployment is comparable to the energy needed to power millions of U.S. homes, reflecting the enormous computing requirements behind the next generation of AI systems.</p>



<p>Kress noted that Nvidia is actively working with OpenAI but emphasized that several elements of the agreement remain under negotiation.</p>



<p>The company has not disclosed further details about timelines or structural terms, maintaining a conservative tone around expectations for when the deal might be finalized.</p>



<p>OpenAI, which accelerated global interest in generative AI with the launch of ChatGPT in 2022, remains one of Nvidia’s most significant customers.</p>



<p>Its demand for high-performance chips has grown along with the rising number of companies building AI-driven systems that rely on large-scale computing clusters powered by Nvidia hardware.</p>



<p>Nvidia has previously confirmed that it has around $500 billion in chip bookings through 2026, reflecting escalating industry demand for advanced GPUs and AI-focused accelerators.</p>



<p>However, Kress stated that any eventual commitments tied to the OpenAI agreement are not yet included in that figure, suggesting potential for substantial additional orders if the partnership is finalized.</p>



<p>She noted that none of the future OpenAI allocations are part of the current half-trillion-dollar forecast, underscoring the potential scale of future demand linked to the deal.</p>



<p>Investors responded positively, with Nvidia’s share price rising during the session following the remarks.</p>



<p>The chipmaker has been expanding its involvement across the AI startup ecosystem over the past year, supporting new players and forming partnerships aimed at accelerating AI development across industries.</p>



<p>This has also led to concerns among some analysts about the risk of circular financing, where companies simultaneously supply, invest in and depend on the same partners for revenue.</p>



<p>Nvidia recently announced plans to commit up to $10 billion to Anthropic, another major player in the AI sector and a direct competitor to OpenAI.</p>



<p>That investment, too, could meaningfully expand Nvidia’s future bookings, further reinforcing the company’s role at the center of the rapidly scaling AI infrastructure supply chain.</p>



<p>Industry observers say the chipmaker’s rising influence reflects the central position of high-performance GPUs in modern AI development.</p>



<p>As companies seek greater computing capacity to train and deploy increasingly complex models, partnerships with hardware providers have become essential to scaling.</p>



<p>While the proposed $100 billion OpenAI agreement has generated intense public interest, Nvidia’s cautious stance suggests that many variables remain under evaluation.</p>



<p>Finalizing such a deal would not only cement a high-profile alliance but could also reshape competition across the AI ecosystem as companies race to secure long-term access to advanced processing power.</p>



<p>For now, Nvidia continues to indicate strong demand across the sector and growing orders from major cloud providers and AI developers.</p>



<p>Its ongoing negotiations with OpenAI highlight the evolving dynamics of the industry, where multibillion-dollar technology partnerships are becoming critical to meeting global expectations for the next generation of artificial intelligence.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>AMD, Cisco and Humain Launch New AI Data Center Venture, Secure First Major Customer</title>
		<link>https://www.millichronicle.com/2025/11/59516.html</link>
		
		<dc:creator><![CDATA[NewsDesk Milli Chronicle]]></dc:creator>
		<pubDate>Wed, 19 Nov 2025 20:25:25 +0000</pubDate>
				<category><![CDATA[Featured]]></category>
		<category><![CDATA[News]]></category>
		<category><![CDATA[World]]></category>
		<category><![CDATA[100 MW data center project]]></category>
		<category><![CDATA[2030 data center expansion]]></category>
		<category><![CDATA[advanced computing clusters]]></category>
		<category><![CDATA[AI data center]]></category>
		<category><![CDATA[AI infrastructure growth]]></category>
		<category><![CDATA[AMD AI chips]]></category>
		<category><![CDATA[Cisco networking]]></category>
		<category><![CDATA[global AI market]]></category>
		<category><![CDATA[high-performance computing Middle East]]></category>
		<category><![CDATA[Humain Saudi startup]]></category>
		<category><![CDATA[large-scale generative AI]]></category>
		<category><![CDATA[Luma AI customer]]></category>
		<category><![CDATA[MI450 accelerators]]></category>
		<category><![CDATA[Middle East technology]]></category>
		<category><![CDATA[renewable energy data centers]]></category>
		<guid isPermaLink="false">https://millichronicle.com/?p=59516</guid>

					<description><![CDATA[AMD , Cisco and Saudi-based Humain unveil a major AI-focused joint venture, beginning with a 100 MW data center and]]></description>
										<content:encoded><![CDATA[
<blockquote class="wp-block-quote">
<p>AMD , Cisco and Saudi-based Humain unveil a major AI-focused joint venture, beginning with a 100 MW data center and landing a flagship customer before construction even begins.</p>
</blockquote>



<p>A new artificial intelligence joint venture is taking shape in the Middle East, as AMD, Cisco and Saudi startup Humain announce plans to build advanced data centers across the region.</p>



<p>The partnership will begin with a 100-megawatt data center in Saudi Arabia, marking the first phase of an ambitious long-term strategy focused on high-performance AI computing.</p>



<p>Humain has already secured the project’s first major customer, with generative video company Luma AI contracting to take the entire 100-megawatt capacity.</p>



<p>Executives involved in the venture confirmed that this full-capacity commitment represents one of the earliest large-scale generative AI compute deals in the region.</p>



<p>The creation of the joint venture follows a surge of U.S.–Saudi technology agreements announced during recent high-level political visits and expanding economic discussions.</p>



<p>Humain, backed by Saudi Arabia’s sovereign wealth fund, has been positioning itself as a major player in regional AI infrastructure development.</p>



<p>The country’s access to vast land resources and competitively priced energy has made it an attractive location for large-scale data center investments.</p>



<p>AMD previously announced a $10 billion collaboration with Humain earlier in the year, covering purchases of advanced AI chips and deepening the companies’ technological ties.</p>



<p>In the new venture, AMD and Cisco will serve as minority shareholders, sharing responsibility for both profits and operational outcomes.</p>



<p>Humain will lead development, construction and long-term planning, with AMD CEO Lisa Su noting that all partners will work collectively to ensure success.</p>



<p>The companies have not released additional financial details about the venture, but indicated that the long-term target is an expansive, multi-country AI compute ecosystem.</p>



<p>The joint venture aims to serve a vast regional market spanning Asia, Europe, India, the Middle East and Africa—reaching a population of nearly 4.5 billion.</p>



<p>Executives outlined a roadmap to build up to one gigawatt of new data center capacity by 2030, with multiple sites planned over the next several years.</p>



<p>The initial 100-megawatt center is planned for completion in 2026, and will rely entirely on renewable energy sources for operation.</p>



<p>Cisco will provide networking systems, hardware and infrastructure support for the project, ensuring connectivity at large scale for the high-load AI computing environment.</p>



<p>AMD will supply its MI450 AI accelerators for the first buildout, offering the processing performance needed for next-generation machine learning tasks.</p>



<p>Humain is already receiving purchase orders for future expansions, indicating strong early demand for high-capacity compute clusters in the region.</p>



<p>Construction has not yet started, but preparatory planning is underway, with teams focused on design, engineering, and energy integration.</p>



<p>Cisco will also leverage its extensive global sales network to help sell future data center capacity across multiple continents.</p>



<p>The company’s executives emphasized their long history of building sales incentivization structures, aiming to accelerate demand for Humain’s upcoming facilities.</p>



<p>The venture is emerging at a time when global competition for AI infrastructure is intensifying, as governments and corporations invest heavily in computing power and next-generation capabilities.</p>



<p>Large-scale generative AI models require vast energy and processing resources, driving new interest in partnerships that combine hardware, networking and regional support.</p>



<p>The commitment from Luma AI to secure the entire output of the first center signals confidence in the project’s long-term potential and operational scale.</p>



<p>Industry observers note that the venture reflects broader shifts in global AI strategy, with the Middle East positioning itself as a rising hub for high-density compute development.</p>



<p>As planning continues, the companies aim to create a system that supports global AI growth, while advancing energy efficiency, economic diversification and emerging digital industries.</p>



<p>The joint venture may become one of the most significant AI data center collaborations of the decade, linking U.S. technology leaders with Saudi Arabia’s rapidly expanding digital infrastructure ambitions.</p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
