VIDEO

Driving Faster Time to Production for AI Inference

Moving AI from architecture to production is where real value is created. In this episode, Charlie and Nilesh discuss how proven reference architectures and strong field collaboration help enterprises deploy AI inference faster and more reliably.

They explain why simplifying deployment, reducing complexity, and accelerating time to production are critical as organizations scale AI factories and bring inference workloads into real-world environments, based on joint experience delivering AI infrastructure with NVIDIA and WEKA.

Transcript

00:06

Nilesh Patel: Hi, everyone. This is Nilesh Patel, Chief Strategy Officer at WEKA. I’m glad to be here with Charlie Boyle, VP of DGX Systems at NVIDIA. Charlie, welcome and thanks for joining.

00:15

Charlie Boyle: Thanks, Nilesh.

00:18

Nilesh Patel: And we are here today discussing some of the key factors that a lot of enterprises have to consider in designing, building, and scaling their AI infrastructures, and take into account some of the challenges of data infrastructure that are associated with that.

One other interesting learning that we have had throughout our experience together is not only designing and building the stack, but also seeing it through deployed. And that’s where our partnership between the two companies, not at a product and technical level that we are having, but also in the field with NVIS and WEKA Pro Serv and how they are achieving and helping customers minimize the time to the outcome, time to the production is extremely important.

00:59

Charlie Boyle: Yeah, totally. I mean, that’s part of the reason that we developed the SuperPOD product years ago was for this exact customer requirement is enterprise customers knew something about AI, they had an idea they wanted to build a center of excellence, but at the end of the day, they just want something to work.

And so everything that we do with SuperPOD and our great partners like WEKA, it’s not just an NVIDIA thing. We make that available to all of our partners because we want everyone deploying NVIDIA AI factories to be successful, whether it’s on DGX, one of our great partner solutions. So, not only building the reference architecture and building that technical foundation that we know how to deploy something, but when you go to deploy in a customer data center, they’re all a little different. Everything’s a little bit unique.

But when you start with that great foundation, it’s easy to adapt in the field to a different cable length, a different power type, all of those things. That’s why it’s so important as our customers are working with NVIDIA and great partners like WEKA, that they understand that we’re deploying a complex solution, but to them it should just be easy. And that joint partnership really enables us to plan, deploy, execute, and then really see customer results very quickly.

02:15

Nilesh Patel: That is so great. And I think we can’t wait to see how we continue to expand and help customers be more successful. I look forward to continuing to partner with you. I think we have tremendous value proposition that is being loaded up to be delivered from our products and technology innovations we are driving together.

Thank you so much for your time today.

02:38

Nilesh Patel: If you’re expanding your AI factories or building new ones, come see how NVIDIA and WEKA are building AI factories together, and let us share our learnings with you.

01:58

Nilesh Patel: If you’re expanding your AI factories or building new ones, come see how NVIDIA and WEKA are building AI factories together and let us share our learnings with you.

Related Resources