ClouderaNOW   Learn about the latest innovations in data, analytics, and AI   |  July 16

Register now
| Partners

Cloudera Supercharges Your Private AI with Cloudera AI Inference, AI-Q NVIDIA Blueprint, and NVIDIA NIM

Zoram Thanga Headshot
Dennis Duckworth headshot
AI

As we speak with our customers about their goals for AI, a common pain point we hear is that their plans and implementations are sometimes stalled due to concerns about privacy. They want to use AI on all of their corporate data since that is the way their employees and customers will get the most accurate results and answers, but they realize they can’t send their data out to a public endpoint for a closed-source large language model (LLM) since, 1) there is too much data, and 2) their data would no longer be private.

To address these concerns, Cloudera has begun espousing the concept of Private AI, which would allow these customers to get all of the benefits that AI brings and keep their proprietary data safe and secure.

NVIDIA is seeing the same challenge, but at a much higher and broader level: nation states. Governments are realizing that it isn’t in the best interests of their nations to run AI in another country, so they’re working to build out the infrastructure that they need to keep their data and their AI within their own borders. They can then control what other countries or entities they share their data or AI results with.

At the GTC Paris conference today, NVIDIA provided the building blocks for Sovereign AI to support governments in their efforts. This initiative aligns well with Cloudera’s focus on enabling our customers to implement their own Private AI platforms. 

NVIDIA made two other announcements that are of particular interest to Cloudera, and in this blog we’ll dive into AI-Q NVIDIA Blueprint for Enterprise Research and the NVIDIA NIM and what this means for our customers.

AI-Q NVIDIA Blueprint with Cloudera AI

NVIDIA’s introduction of the AI-Q blueprint for enterprise research provides Cloudera AI with more capabilities for supporting our customers’ complex agentic AI needs. 

Cloudera AI Inference can host all of the NVIDIA NeMo Retriever and LLM inference microservices that make up the AI-Q NVIDIA Blueprint, including NVIDIA Llama Nemotron reasoning models. Combining the strong privacy and security provided by the Cloudera AI platform for the model endpoints with the powerful NVIDIA Agent Intelligence toolkit, you can take your enterprise agentic applications to the next level.

Benefits of Using AI-Q NVIDIA Blueprint with Cloudera AI
Leveraging AI-Q NVIDIA Blueprint within Cloudera AI Inference service unlocks massive  AI potential. This powerful combination integrates leading reasoning models packaged as NVIDIA NIM and NeMo Retriever microservices onto Cloudera AI, and it ensures seamless connectivity between agents, tools, and data through full compatibility with the NVIDIA Agent Intelligence toolkit. 

This multi-framework capability empowers organizations to build sophisticated enterprise retrieval-augmented generation (RAG) applications with robust privacy and security, taking full advantage of state-of-the-art AI advancements.

NVIDIA NIM  with Cloudera AI Inference

NVIDIA's NIM container is a game-changer for getting the best performance from LLMs quickly and easily: it significantly speeds up LLM deployment and inference by automatically selecting the best inference backend based on the model and GPU hardware, enabling a model-agnostic inference solution that streamlines the production serving of numerous cutting-edge LLMs. 

Digging deeper, the universal LLM NIM microservice enables users to quickly deploy LLMs accelerated by NVIDIA TensorRT-LLM, vLLM, or SGLang for top-tier inference on any NVIDIA accelerated platform. It supports models stored in Hugging Face or TensorRT-LLM formats, enabling enterprise-grade inference for a vast array of LLMs. Users can rely on smart defaults for optimized latency and throughput or fine-tune performance with simple configuration options. As part of NVIDIA AI Enterprise, the universal LLM NIM microservice receives continuous updates from NVIDIA, ensuring compatibility with a wide range of popular LLMs.

Benefits of Using the NVIDIA NIM within Cloudera AI Inference
NVIDIA's NIM provides our customers more flexibility in how they can make use of LLMs in their AI applications. Cloudera AI Inference service already has NVIDIA NIM embedded in it, so customers can implement the universal LLM NIM quickly and easily. Customers get the benefits of NVIDIA NIM with the ease of use, security, and streamlined support of a single, unified platform: Cloudera.

Through its seamless integration into our AI Inference service, the universal LLM NIM microservice offers significant advantages for Cloudera AI customers, including:

  • Accelerated deployment: Get your LLM applications up and running faster with pre-built, optimized containers.

  • Enhanced performance: Leverage the full potential of NVIDIA accelerated computing for high-speed inference and reduced latency.

  • Scalability: Easily scale your LLM deployments to meet the demands of your growing business.

  • Simplified management: Manage and monitor your LLM deployments with Cloudera's intuitive interface.

Conclusion

Together, Cloudera and NVIDIA empower businesses to leverage the latest advancements in AI easily, efficiently, and cost-effectively on all of their data, whether public or private. By simplifying the AI application lifecycle, from development to deployment, and by optimizing performance, we're helping our users unlock the full potential of AI.

Be sure to check out NVIDIA’s blog about announcements out of GTC Paris and  Cloudera’s blogs on AI, especially the most recent one about “AI in a Box,” powered by Dell, NVIDIA, and Cloudera  which gives customers a  new way to implement  Private AI quickly, easily, and with minimal risk. 

 

Ready to Get Started?

Your form submission has failed.

This may have been caused by one of the following:

  • Your request timed out
  • A plugin/browser extension blocked the submission. If you have an ad blocking plugin please disable it and close this message to reload the page.