RIEM News LogoRIEM News

Articles tagged with "GPUs"

  • Nvidia bulks up open source offerings with an acquisition and new open AI models 

    Nvidia is strengthening its presence in open source AI through two major initiatives: the acquisition of SchedMD and the release of a new family of open AI models. SchedMD, founded in 2010 by the original developers of the widely used open source workload management system Slurm, has been a long-term partner of Nvidia. The acquisition, with undisclosed terms, aims to leverage SchedMD’s technology as critical infrastructure for generative AI, enabling Nvidia to accelerate access to diverse computing systems. Nvidia plans to continue investing in this technology to support AI development at scale. In addition to the acquisition, Nvidia introduced the Nemotron family of open AI models, which it claims to be the most efficient open models for building accurate AI agents. This lineup includes the Nemotron 3 Nano for targeted tasks, Nemotron 3 Super for multi-agent AI applications, and Nemotron 3 Ultra for more complex tasks. Nvidia’s CEO Jensen Huang emphasized that Nemotron represents a move toward open innovation,

    robotAI-modelsNvidiaopen-source-AIgenerative-AIworkload-managementGPUs
  • Microsoft inks $9.7B deal with Australia’s IREN for AI cloud capacity

    Microsoft has secured a significant $9.7 billion, five-year contract with Australia-based IREN to expand its AI cloud computing capacity. This deal grants Microsoft access to advanced compute infrastructure equipped with Nvidia GB300 GPUs, which will be deployed in phases through 2026 at IREN’s facility in Childress, Texas, designed to support up to 750 megawatts of capacity. Separately, IREN is investing about $5.8 billion in GPUs and equipment from Dell to support this infrastructure expansion. The agreement follows Microsoft’s recent launch of AI models optimized for reasoning, agentic AI systems, and multi-modal generative AI, reflecting the company's efforts to meet growing demand for AI services. Microsoft has also previously acquired approximately 200,000 Nvidia GB300 GPUs for data centers in Europe and the U.S. IREN, originally a bitcoin-mining firm, has pivoted successfully to AI workloads, leveraging its extensive GPU resources. CEO Daniel Roberts anticipates that the Microsoft contract will utilize only

    energycloud-computingAI-infrastructureGPUsdata-centersMicrosoftNvidia
  • Nvidia becomes first public company worth $5 trillion

    Nvidia has become the first public company to reach a $5 trillion market capitalization, driven primarily by its dominant position in the AI chip market. The company’s shares surged over 5.6% following news that U.S. President Donald Trump planned to discuss Nvidia’s Blackwell chips with Chinese President Xi Jinping. Nvidia CEO Jensen Huang highlighted the company’s expectation of $500 billion in AI chip sales and emphasized expansion into sectors such as security, energy, and science, which will require thousands of Nvidia GPUs. Additionally, Nvidia is investing in enabling AI-native 5G-Advanced and 6G networks through its platforms, further solidifying its role in the AI infrastructure ecosystem. This milestone comes just three months after Nvidia first surpassed a $1 trillion valuation, with its stock rising more than 50% in 2025 due to strong demand for its GPUs used in data centers for training large language models and AI inference. Nvidia’s GPUs remain scarce and highly sought after, supporting the growing infrastructure needed

    energyAI-chipsGPUsdata-centersNvidia5G-networks6G-networks
  • Nscale inks massive AI infrastructure deal with Microsoft

    Nscale, an AI cloud provider founded in 2024, has secured a major deal to deploy approximately 200,000 Nvidia GB300 GPUs across data centers in Europe and the U.S. This deployment will occur through Nscale’s own operations and a joint venture with investor Aker. Key locations include a Texas data center leased by Ionic Digital, which will receive 104,000 GPUs over 12 to 18 months, with plans to expand capacity to 1.2 gigawatts. Additional deployments include 12,600 GPUs at the Start Campus in Sines, Portugal (starting Q1 2026), 23,000 GPUs at Nscale’s Loughton, England campus (starting 2027), and 52,000 GPUs at Microsoft’s AI campus in Narvik, Norway. This deal builds on prior collaborations with Microsoft and Aker involving data centers in Norway and the UK. Josh Payne, Nscale’s founder and CEO, emphasized that this agreement positions Nscale as

    energyAI-infrastructuredata-centersGPUssustainabilitycloud-computingtechnology-investment
  • While OpenAI races to build AI data centers, Nadella reminds us that Microsoft already has them

    Microsoft CEO Satya Nadella announced the deployment of the company’s first massive AI system—referred to as an AI “factory” by Nvidia—at Microsoft Azure’s global data centers. These systems consist of clusters with over 4,600 Nvidia GB300 rack computers equipped with the new Blackwell Ultra GPU chips, connected via Nvidia’s high-speed InfiniBand networking technology. Microsoft plans to deploy hundreds of thousands of these Blackwell Ultra GPUs worldwide, enabling the company to run advanced AI workloads, including those from its partner OpenAI. This announcement comes shortly after OpenAI secured significant data center deals and committed approximately $1 trillion in 2025 to build its own infrastructure. Microsoft emphasized that, unlike OpenAI’s ongoing build-out, it already operates extensive data centers in 34 countries, positioning itself as uniquely capable of supporting frontier AI demands today. The new AI systems are designed to handle next-generation AI models with hundreds of trillions of parameters. Further details on Microsoft’s AI infrastructure expansion are

    energydata-centersAI-hardwareGPUscloud-computingNvidiaMicrosoft-Azure
  • Wall Street analysts explain how AMD’s own stock will pay for OpenAI’s billions in chip purchases 

    AMD and OpenAI have announced an expanded partnership in which OpenAI will assist AMD in refining its Instinct GPUs—AMD’s competitor to Nvidia chips—and commit to purchasing 6 gigawatts of compute capacity over several years. The deal is valued in the billions, but rather than paying with cash, OpenAI will use AMD stock to finance its purchases. AMD has granted OpenAI up to 160 million stock warrants, which vest as certain milestones are met, including significant increases in AMD’s stock price. For example, the final tranche requires AMD’s market cap to reach around $1 trillion, implying a potential value of about $100 billion for OpenAI’s stake if all conditions are met and shares are held without selling. UBS analyst Timothy Arcuri suggests that OpenAI will likely sell portions of its AMD stock over time to cover its GPU purchases, effectively making this a financing arrangement for AMD. Despite the unconventional structure, the deal serves as a strong validation of AMD’s AI GPU capabilities,

    energyAI-chipsAMDOpenAIGPUssemiconductor-materialscompute-capacity
  • 6-gigawatt handshake: AMD joins OpenAI’s trillion-dollar AI plan

    OpenAI has entered a landmark multi-year agreement with AMD to deploy up to 6 gigawatts of AMD Instinct GPUs, marking one of the largest GPU deployment deals in AI history. The partnership will start with a 1-gigawatt rollout of AMD’s upcoming MI450 GPUs in late 2026 and scale to 6 gigawatts over multiple hardware generations, powering OpenAI’s future AI models and services. This collaboration builds on their existing relationship involving AMD’s MI300X and MI350X GPUs, with both companies committing to jointly advance AI hardware and software through shared technical expertise. Following the announcement, AMD’s stock surged nearly 24%, reflecting strong market confidence. A significant component of the deal includes an equity arrangement whereby OpenAI received a warrant for up to 160 million AMD shares, potentially giving OpenAI about a 10% stake in AMD if fully exercised. The warrant vests in stages tied to deployment milestones and AMD’s stock price. Although the exact financial terms

    energyAI-hardwareGPUsAMDOpenAIhigh-performance-computingAI-compute-capacity
  • AMD to supply 6GW of compute capacity to OpenAI in chip deal worth tens of billions

    AMD has entered a multi-year chip supply agreement with OpenAI that could generate tens of billions in revenue and significantly boost AMD’s presence in the AI sector. Under the deal, AMD will provide OpenAI with 6 gigawatts of compute capacity using multiple generations of its Instinct GPUs, beginning with the Instinct MI450 GPU, which is expected to be deployed in the second half of 2026. AMD claims the MI450 will outperform comparable Nvidia GPUs through hardware and software enhancements developed with OpenAI’s collaboration. Currently, OpenAI already uses AMD’s MI355X and MI300X GPUs for AI inference tasks due to their high memory capacity and bandwidth. In addition to supplying chips, AMD has granted OpenAI the option to purchase up to 160 million shares of AMD stock, representing a 10% stake. The stock vesting is tied to the deployment milestones of the compute capacity and AMD’s stock price, with the final tranche vesting if AMD shares reach $600. Following the

    energyAI-computeGPUsdata-centerschip-supplysemiconductorAI-infrastructure
  • NVIDIA investing $100B in OpenAI data centers for next-gen AI

    OpenAI and NVIDIA have entered a landmark partnership, with NVIDIA committing up to $100 billion to build massive AI data centers that will deploy at least 10 gigawatts of compute power using millions of NVIDIA GPUs. The first gigawatt of this capacity is expected to go live in the second half of 2026 on NVIDIA’s upcoming Vera Rubin platform. NVIDIA CEO Jensen Huang described the collaboration as a “next leap forward” for both companies, highlighting that the 10 gigawatts equate to roughly 4 to 5 million GPUs—double the number shipped by NVIDIA last year. This massive infrastructure investment underscores the deep ties between the two companies and their joint efforts to power the next era of AI intelligence. OpenAI CEO Sam Altman emphasized that compute infrastructure is central to OpenAI’s mission and will form the foundation of the future economy. He noted the challenge of balancing research, product development, and scaling infrastructure, promising significant developments in the coming months. OpenAI cofounder Greg

    energydata-centersAI-infrastructureNVIDIAOpenAIGPUscompute-power
  • How a once-tiny research lab helped Nvidia become a $4 trillion-dollar company

    The article chronicles the evolution of Nvidia’s research lab from a small group of about a dozen people in 2009, primarily focused on ray tracing, into a robust team of over 400 researchers that has been instrumental in transforming Nvidia from a video game GPU startup into a $4 trillion company driving the AI revolution. Bill Dally, who joined the lab after being persuaded by Nvidia leadership, expanded the lab’s focus beyond graphics to include circuit design and VLSI chip integration. Early on, the lab recognized the potential of AI and began developing specialized GPUs and software for AI applications well before the current surge in AI demand, positioning Nvidia as a leader in AI hardware. Currently, Nvidia’s research efforts are pivoting toward physical AI and robotics, aiming to develop the core technologies that will power future robots. This shift is exemplified by the work of Sanja Fidler, who joined Nvidia in 2018 to lead the Omniverse research lab in Toronto, focusing on simulation models for robotics and

    robotartificial-intelligenceNvidiaGPUsrobotics-developmentAI-hardwaretechnology-research
  • Two arrested for smuggling AI chips to China; Nvidia says no to kill switches

    The U.S. Department of Justice arrested Chuan Geng and Shiwei Yang on August 2 in California for allegedly smuggling advanced AI chips to China through their company, ALX Solutions. They face charges under the Export Control Reform Act, which carries penalties of up to 20 years in prison. The DOJ indicated the chips involved were highly powerful GPUs designed specifically for AI applications, strongly suggesting Nvidia’s H100 GPUs. Evidence showed ALX Solutions shipped these chips to intermediaries in Singapore and Malaysia while receiving payments from entities in Hong Kong and China, apparently to circumvent U.S. export restrictions. In response, Nvidia emphasized its strict compliance with U.S. export controls and stated that any diverted products would lack service and support. The company also rejected recent U.S. government proposals to embed kill switches or backdoors in chips to prevent smuggling, arguing such measures would compromise security and trust in U.S. technology. Nvidia warned that creating vulnerabilities intentionally would benefit hackers and hostile actors, ultimately harming America

    AIsemiconductorsNvidiaexport-controlchip-smugglingtechnology-securityGPUs
  • Nvidia Breaks $4 Trillion Market Value Record

    Nvidia has become the first publicly traded company to reach a $4 trillion market valuation, surpassing established tech giants such as Apple, Microsoft, and Google. Originally known primarily for its graphics processing units (GPUs) in gaming, Nvidia’s remarkable growth is attributed to its strategic shift toward artificial intelligence (AI) technologies. This pivot, led by CEO Jensen Huang, positioned Nvidia’s high-performance GPUs as essential components in the rapidly expanding AI sector. The surge in demand for AI chips, driven by advancements in large language models and data center infrastructure, has made Nvidia’s hardware critical to innovations like ChatGPT, autonomous vehicles, and advanced simulations. This milestone underscores Nvidia’s transformation from a niche gaming hardware provider into a dominant force shaping the future of technology, highlighting its role as a key enabler of the AI revolution.

    robotAIautonomous-vehiclesGPUsdata-centersartificial-intelligenceNvidia
  • ChatGPT: Everything you need to know about the AI-powered chatbot

    ChatGPT, OpenAI’s AI-powered text-generating chatbot, has rapidly grown since its launch to reach 300 million weekly active users. In 2024, OpenAI made significant strides with new generative AI offerings and the highly anticipated launch of its OpenAI platform, despite facing internal executive departures and legal challenges related to copyright infringement and its shift toward a for-profit model. As of 2025, OpenAI is contending with perceptions of losing ground in the AI race, while working to strengthen ties with Washington and secure one of the largest funding rounds in history. Recent updates in 2025 include OpenAI’s strategic use of Google’s AI chips alongside Nvidia GPUs to power its products, marking a diversification in hardware. A new MIT study raised concerns that ChatGPT usage may impair critical thinking by showing reduced brain engagement compared to traditional writing methods. The ChatGPT iOS app saw 29.6 million downloads in the past month, highlighting its massive popularity. OpenAI also launched o3

    energyartificial-intelligenceOpenAIGPUsAI-chipspower-consumptionmachine-learning