Deepinfra’s $107M Series B: Building a Dedicated Inference Cloud for Open-Source AI
Deepinfra, a startup specializing in dedicated inference cloud services for open-source artificial intelligence models, recently announced a significant milestone: a $107 million Series B funding round led by 500 Global and Georges Harik, an early Google cloud engineer. This investment, which also includes participation from tech giants like Nvidia and Samsung Next, will accelerate the expansion of Deepinfra’s global infrastructure. Below, we dive into the key details about the company, its technology, and what this funding means for the future of open-source AI deployment.
What is Deepinfra and what does it do?
Deepinfra is a cloud startup that provides a dedicated inference cloud specifically designed for running open-source AI models. Unlike general-purpose cloud platforms, Deepinfra’s infrastructure is optimized for the unique demands of model inference—the process of generating predictions or outputs from a trained AI model. The company focuses on open-source models, allowing developers to deploy and scale popular frameworks like Llama, Mistral, and Stable Diffusion with minimal latency and high throughput. By offering dedicated compute resources, Deepinfra ensures that customers avoid “noisy neighbor” issues common in shared cloud environments, resulting in more consistent performance for production AI workloads.

How much funding did Deepinfra raise and who led the round?
Deepinfra raised $107 million in a Series B funding round. The round was co-led by 500 Global, a well-known venture capital firm, and Georges Harik, one of Google’s first cloud engineers. Harik’s involvement brings deep expertise in cloud infrastructure, which aligns with Deepinfra’s mission to build a high-performance inference cloud. This funding represents a strong vote of confidence in the company’s technology and market position, especially given the growing demand for efficient AI inference at scale.
Who participated in the Series B funding round?
In addition to the lead investors, the Series B round saw participation from several heavy hitters in the technology and semiconductor industries. Notably, Nvidia Corp. joined as an investor, underscoring the close relationship between GPU hardware and AI inference workloads. Other participants include Samsung Next, the investment arm of Samsung, and several unnamed strategic partners. The involvement of Nvidia is particularly significant, as it signals a deeper integration between Deepinfra’s cloud services and Nvidia’s GPU platforms, potentially giving Deepinfra early access to next-generation hardware for inference.
What is a dedicated inference cloud and why is it important for open-source models?
A dedicated inference cloud is a cloud infrastructure purpose-built for running AI model inference tasks, as opposed to general-purpose cloud services that handle a variety of workloads. For open-source models, this is crucial because inference often requires specialized hardware like GPUs or TPUs, low-latency networking, and efficient memory management to handle large model sizes. Deepinfra’s dedicated approach means that customers get consistent performance without competition for resources. This is especially important for production applications such as chatbots, image generators, and code assistants, where even small delays can degrade user experience. By focusing exclusively on inference, Deepinfra can optimize every layer of the stack—from the physical servers to the software load balancers—for the specific demands of open-source AI models.

How does Deepinfra’s service differ from general-purpose cloud providers?
Unlike general-purpose cloud providers like AWS, Azure, or Google Cloud, Deepinfra is purpose-built for AI inference. This specialization allows the company to offer lower latency, better cost efficiency, and simpler deployment for open-source models. General-purpose clouds often require users to manually configure GPU instances, set up networking, and manage scaling, which can be complex and error-prone. Deepinfra abstracts away these complexities by providing a pay-as-you-go API that developers can integrate in minutes. Additionally, because Deepinfra aggregates demand across many customers, it can achieve higher utilization rates, passing on cost savings. The dedicated infrastructure also eliminates the “cold start” problem in serverless environments, ensuring that models are always warm and ready to serve requests.
What are the plans for the $107 million funding?
Deepinfra plans to use the $107 million to expand its global infrastructure, adding more data center locations to reduce latency for users worldwide. The company also intends to invest in research and development to improve its inference engine, support for more open-source model architectures, and better tooling for developers. Part of the funding will go toward hiring engineering talent and building out its sales and marketing teams. Given the participation of Nvidia, Deepinfra may also forge closer technical collaborations to optimize its platform for future GPU generations. Ultimately, the goal is to make open-source AI models as accessible and performant as proprietary alternatives, accelerating adoption across industries.
Related Articles
- The Enduring Value of Developer Communities in an AI Era
- Nebius Boosts AI Infrastructure with $643M Acquisition of Eigen AI
- AI-Driven Care: The Hidden Risks of Removing Human Touch from Healthcare
- 7 Strategies for Bootstrapping Success: Lessons from Runpod's Community-Funded Rise
- Remote Work 'Career Suicide,' Warns Good American Co-Founder Emma Grede — Cites Loneliness Epidemic
- The Enduring Power of Developer Communities in the AI Era
- How to Build AI Applications Without Relying on Collapsing Scaffolding: A Step-by-Step Guide
- 10 Essential Principles for Building Trust Before Monetizing Attention as a Founder