Introducing GMI Cloud: New On-Demand Instances Accelerate Access to NVIDIA GPUs
GMI Cloud, the emerging GPU cloud platform designed for AI and ML workloads, is accelerating access to NVIDIA GPUs. Its new On-Demand cloud compute offering, available today, is built for companies that are serious about leveraging AI and unlocking the door from prototyping to production. Users can access GMI Cloudâs On-Demand GPU computing resources almost instantaneously.
The Surge in Demand for Compute
The current surge in demand for AI compute power requires companies to be strategic in their approach. In a fast-evolving landscape, organizations are being asked to pay a 25-50% down payment and sign up for a 3-year contract with the promise of getting access to GPU infrastructure in 6-12 months. The AI shift has forced companies to need more flexible computing power.
Instant GPUs, Infinite AI
Leveraging its ties to Realtek Semiconductors (TPE: 2379) and GMI Technologies (TPE: 3312) as well as Taiwan’s robust supply chain ecosystem, GMI Cloud is able to ensure quicker deployment and higher efficiency in operations. The physical presence in Taiwan cuts the GPU delivery time down from months to days compared to non-Taiwanese GPU providers. GMI Cloud is poised to become the most competitive new entrant in this market.
âOur mission is to empower humanity’s AI ambitions with instant, efficient GPU cloud,â said Alex Yeh, Founder and CEO of GMI Cloud. âWe’re not just building a cloudâwe’re creating the backbone of the AI era. GMI Cloud is dedicated to transforming how developers and data scientists leverage NVIDIA GPUs and how all humans benefit from AI.â
Why It Matters
Technology leaders are seizing the opportunities presented by the growing AI wave, and organizations of all sizes are hitting walls when it comes to accessing compute power.
Startups, for example, donât have the budget or long-term forecasting to pay a down payment for a large GPU installation. They need the flexibility to scale up or down based on their traction. This requires the option to pay for GPU as an operating expense rather than locking in capital that could be spent on hiring competitive AI talent. On-Demand access provides an instant, cost-effective, and scalable option for teams that need access to GPU compute, without requiring special skills to set up the infrastructure.
Large enterprises face hurdles as well. Enterprise data science teams, for example, require the flexibility to experiment, prototype, and evaluate AI applications to get ahead of competitors before the AI wave passes them by. However, not every enterprise is ready to commit to the long-term contracts and unproven capital expenditures required for larger compute reserves. The flexibility of instant GPU access allows those data science teams to run several prototyping projects that require processing large datasets or fine-tuning models without taking significant investment risks.
Get Started Now
GMI Cloud is a GPU cloud platform, powered by NVIDIA, with a rich Kubernetes-managed preloaded software stack designed specifically for AI and ML workloads which includes prebuilt images with NVIDIA TensorRT and will soon support all NVIDIA prebuilt containers including inference servers like NVIDIA Triton. With competitive pricing offered at $4.39/hour for NVIDIA H100 Tensor Core GPUs, GMI Cloud offers affordable on-demand access compared to larger organizations charging up to 4X that cost for on-demand in an effort to lock users into larger reserve packages. The instance type and size is purposefully designed to efficiently deploy, fine-tune, and inference models ranging in size from Llama3 8b, 70b, Mixtral 8x7b, Google Gemma, Stable Diffusion, and more.