Season 1 | Episode 11

Inference, Guardrails, and Observability for LLMs with Jonathan Cohen

In this episode of AI Explained, we are joined by Jonathan Cohen, VP of Applied Research at NVIDIA. 

We will explore the intricacies of NVIDIA's NeMo platform and its components like NeMo Guardrails and NIMS. Jonathan explains how these tools help in deploying and managing AI models with a focus on observability, security, and efficiency. They also explore topics such as the evolving role of AI agents, the importance of guardrails in maintaining responsible AI, and real-world examples of successful AI deployments in enterprises like Amdocs. Listeners will gain insights into NVIDIA's AI strategy and the practical aspects of deploying large language models in various industries.

About the guest
Jonathan Cohen is a VP of Applied Research at NVIDIA, where he is the engineering leader of the Nemo platform. He focuses on incubating new AI technology into products, including NIMs, BioNemo, LLM alignment, speech language models, Nemo guardrails, foundation models for human biology, and genomics analysis. Jonathan spent a total of 14 years at NVIDIA in a variety of engineering and research roles, with a three-year stint at Apple as Director of Engineering in the middle. Earlier in his career, he specialized in computer graphics in the visual effect industry, winning a Scientific and Technical Academy Award in 2007.
Transcript
Subscribe