AI Inferencing

  • Boosting AI Data Governance Through Disconnected Clouds

    Microsoft is enhancing cloud computing with sovereign private cloud solutions for businesses, especially in regulated industries. These offerings enable robust data governance and operational continuity, even in fully disconnected environments. The integrated Azure, Microsoft 365, and Foundry Local architecture supports consistent, resilient experiences. Foundry Local now allows offline AI inferencing with large language models, ensuring data remains within customer-controlled perimeters. This innovation empowers organizations with digital sovereignty and advanced capabilities, regardless of connectivity.

    2026年2月24日
  • Oracle Cloud to Deploy 50,000 AMD AI Chips, Challenging Nvidia’s Dominance

    Oracle Cloud Infrastructure will deploy 50,000 AMD GPUs by late 2026, intensifying competition in cloud AI. AMD shares rose slightly on the news, while Oracle and Nvidia dipped. This reflects a trend of cloud providers diversifying GPU options beyond Nvidia, especially for AI inferencing. The partnership utilizes AMD’s Instinct MI450 chips for scalable AI computation. OpenAI is expanding its AMD infrastructure with a multi-year deal, potentially acquiring AMD shares. They also signed a $300 billion cloud agreement with Oracle, showing broader investment in diverse computing resources.

    2025年10月17日
  • Alibaba’s New AI Chip: What We Know

    Alibaba is developing a new AI inferencing chip amidst US-China trade tensions and restrictions on Nvidia exports. This move is part of a broader effort by Chinese tech companies to create domestic chip solutions. Alibaba will use the chip to enhance its cloud infrastructure, offering computing power to customers. This complements, rather than replaces, its reliance on vendors like Nvidia. Fueled by AI-related products, Alibaba’s cloud division saw significant revenue growth. Other Chinese firms, including Baidu and Huawei, are also developing proprietary semiconductors.

    2025年8月29日