NVIDIA and Fiddler: Safe and Trustworthy LLM Applications

Harness the combined power of NVIDIA NIM, NeMo Guardrails, and Fiddler to deploy, monitor and safeguard LLM applications.

Scalable LLM Applications with NVIDIA and Fiddler

Deploying LLM applications at scale presents challenges, from handling complex inferences to ensuring trust and compliance. Without the right generative AI (GenAI) infrastructure, performance bottlenecks, unsafe content, and compliance issues can derail the most promising GenAI initiatives from going into production. 
By combining NVIDIA’s powerful AI infrastructure — including NVIDIA Inference Microservices (NIM) and NVIDIA NeMo Guardrails — with the Fiddler AI Observability Platform, enterprises can now operate LLM applications across diverse use cases with confidence and precision.
Diagram showcasing the integration of application code, NVIDIA NIM, NVIDIA NeMo Guardrails, and LLM, with logs published to the Fiddler AI Observability Platform. The Fiddler dashboard displays graphs for monitoring rail executions, jailbreak attempts, and root cause analysis for LLM improvement.
NVIDIA NIM dashboard featuring charts such as Total Cost Tracker, Cyber Risk Profile, Prompt UMAP, and Jailbreak Tracker.

Productionize Enterprise LLM Applications with NVIDIA NIM and Fiddler

NVIDIA NIM accelerates the deployment of AI foundation models by offering portable, quick inference capabilities with minimal setup. Its optimized microservices enable enterprises to deploy powerful generative AI models at scale, delivering low-latency, high-throughput inferences through engines like TensorRT and TensorRT-LLM. By packaging AI models into containers, NIM simplifies integration into existing Generative AI infrastructure with minimal code changes.

When integrated with the Fiddler AI Observability Platform, NVIDIA NIM ensures that LLM deployments scale efficiently while remaining secure and trustworthy. Prompts processed by NIM are logged and automatically routed to Fiddler’s platform, enabling real-time guardrailing and monitoring, actionable insights, and enhanced operational oversight within a comprehensive Generative AI infrastructure.

Ensure LLM Safety with NVIDIA NeMo Guardrails and Fiddler

NeMo Guardrails provides developers with a rich set of rails to moderate and protect conversations. Developers can define the behavior of LLM applications on specific topics and steer LLMs along pre-defined conversational paths for trustworthy and secure dialog.

Once integrated with Fiddler, the prompts, responses, metadata, and rails executed in NeMo Guardrails are published to the Fiddler AI Observability platform. This enables AI and app engineers to:

  • Gain insights into rails execution
  • Define rich alerts for rail activities
  • Detect hallucination, drift, safety, and operational issues

Fiddler’s custom dashboards and reports provide developers with deep insights into LLM and rail metrics, enabling root cause analysis and effective issue resolution.

Visualize rails executed by NVIDIA NeMo Guardrails in the Fiddler AI Observability Platform with graph displaying PII and Toxicity Rail activation occurrences over time with traffic data. Two lines represent PII and Toxicity activations, highlighting spikes and trends.

Frequently Asked Questions

How do I purchase Fiddler on AWS?

Fiddler is a featured partner with a native integration to SageMaker AI. You can leverage AWS credits and procurement via the Amazon SageMaker AI, enterprises can purchase the Fiddler AI Observability platform without lengthy approval processes. 

Purchasing through the Amazon SageMaker AI provides a native Fiddler and SageMaker AI integration, ensuring a seamless experience within the AWS ecosystem. It eliminates extra security hurdles and InfoSec reviews, speeding up production ML deployment at scale.

What Fiddler capabilities do I get when purchasing Fiddler through Amazon SageMaker AI?

Fiddler can support ML models (computer vision, NLP, and complex multimodal deep learning models), including alerts, customizable metrics, dashboards and reports (e.g. time series and analytics charts, feature impact insights) and root cause analysis capabilities.

You can contact us if you need a custom solution for specific use cases or LLM monitoring capabilities.