SoftBank Corp. announced that its Infrinia Team1-focused on advancing next-generation AI infrastructure architecture and systems-has developed “Infrinia AI Cloud OS,” a comprehensive software stack2 purpose-built for AI data centers.
With the deployment of “Infrinia AI Cloud OS,” AI data center operators can offer Kubernetes*3 as a Service (KaaS) within a secure, multi-tenant environment, as well as Inference as a Service (Inf-aaS) that delivers large language model (LLM) inference through APIs. These capabilities can be integrated directly into operators’ own GPU cloud services. Compared with bespoke architectures or in-house development, the software stack is designed to significantly reduce total cost of ownership (TCO) and operational complexity. This enables faster rollout of GPU cloud services that efficiently and flexibly support the entire AI lifecycle-from large-scale model training through real-time inference.
SoftBank plans to initially deploy “Infrinia AI Cloud OS” within its own GPU cloud offerings. Looking ahead, the Infrinia Team intends to extend deployment to overseas data centers and cloud environments, positioning the platform for global adoption.
Also Read: IBM Unveils Enterprise Advantage for Businesses Scale Agentic AI
Background on the Development of “Infrinia AI Cloud OS”
Demand for GPU-accelerated AI computing is growing rapidly across a wide range of sectors, including generative AI, autonomous robotics, simulation, drug discovery, and advanced materials development. As AI workloads continue to diversify and mature, user requirements for AI infrastructure are becoming more sophisticated, driving demand for capabilities such as:
-
Access to fully managed infrastructure provided by GPU cloud service operators, abstracting GPU bare-metal servers
-
Cost-efficient, highly abstracted inference services that eliminate the need for direct GPU management
-
Advanced operational models where AI systems are trained and optimized in centralized environments and then deployed for inference at the edge
Building and operating GPU cloud services that meet these evolving requirements requires deep technical expertise and involves complex, resource-intensive operations—creating a significant burden for cloud service providers.
To address these challenges, SoftBank’s Infrinia Team developed “Infrinia AI Cloud OS,” a unified software stack designed to maximize GPU performance while enabling rapid deployment and streamlined operation of advanced GPU cloud services.
“To further deepen the utilization of AI as it evolves toward AI agents and Physical AI, SoftBank is launching a new GPU cloud service and software business to provide the essential capabilities required for the large-scale deployment of AI in society. At the core of this initiative is our in-house developed ‘Infrinia AI Cloud OS,’ a GPU cloud platform software designed for next-generation AI infrastructure that seamlessly connects AI data centers, enterprises, service providers and developers. The advancement of AI infrastructure requires not only physical components such as GPU servers and storage, but also software that integrates these resources and enables them to be delivered flexibly and at scale. Through Infrinia, SoftBank will play a central role in building the cloud foundation for the AI era and delivering sustainable value to society.” said, Junichi Miyakawa, President & CEO of SoftBank Corp.






















