MLOps Community
+00:00 GMT

Collections

All Collections

Raise Summit AI Conversations powered by Prosus Group
8 Items

All Content

All
Jure Leskovec
Demetrios Brinkmann
Jure Leskovec & Demetrios Brinkmann · Nov 25th, 2025
Relational Foundation Models: Unlocking the Next Frontier of Enterprise AI
Today’s foundation models excel at text and images—but they miss the relationships that define how the world works. In every enterprise, value emerges from connections: customers to products, suppliers to shipments, molecules to targets. This talk introduces Relational Foundation Models (RFMs)—a new class of models that reason over interactions, not just data points. Drawing on advances in graph neural networks and large-scale ML systems, I’ll show how RFMs capture structure, enable richer reasoning, and deliver measurable business impact. Audience will learn where relational modeling drives the biggest wins, how to build the data backbone for it, and how to operationalize these models responsibly and at scale.
# Structured Data
# Relational Deep Learning
# Enterprise AI
Kopal Garg
Kopal Garg · Nov 25th, 2025
An end-to-end DenseNet-121 pipeline on the MedNIST dataset was rebuilt using NVIDIA’s GPU-native tools, replacing traditional CPU-based stages like Pillow, OpenCV, and PyTorch DataLoader. The GPU workflow delivered 3.3× higher throughput, ~3× lower latency, better memory efficiency, and higher hardware utilization on a Tesla T4. The post also outlines future gains through TensorRT, INT8 quantization, RAPIDS cuDF, and GPUdirect Storage to push medical imaging pipelines closer to real-time performance.
# Inference Optimization
# NVIDIA
# Medical AI
Jeff Huber drops some hard truths about “context rot” — the slow decay of AI memory that’s quietly breaking your favorite models. From retrieval chaos to the hidden limits of context windows, he and Demetrios Brinkmann unpack why most AI systems forget what matters and how Chroma is rethinking the entire retrieval stack. It’s a bold look at whether smarter AI means cleaner context — or just better ways to hide the mess.
# Context Rot
# Search
# AI Agents
# AI Engineering
# Chroma
Brooke Hopkins
Demetrios Brinkmann
Brooke Hopkins & Demetrios Brinkmann · Nov 18th, 2025
Voice AI is finally growing up—but not without drama. Brooke Hopkins joins Demetrios Brinkmann to unpack why most “smart” voice systems still feel dumb, what it actually takes to make them reliable, and how startups are quietly outpacing big tech in building the next generation of voice agents.
# AI Voice Agent
# Voice AI Simulation
# Coval
Combo-Banana is an open-source prototype based on Google's Nano Banana, designed to empower product designers by allowing them to quickly define and automate their multi-step image editing workflows. This tool transforms tedious, repetitive tasks—like background removal, color correction, and integration into new scenes—into a fast, consistent, and automated process, freeing designers to focus on creative work.
# Generative AI Tools
# Productivity
# Artificial Intelligence
# Python
# Open Source
Rani Radhakrishnan
Demetrios Brinkmann
Rani Radhakrishnan & Demetrios Brinkmann · Nov 14th, 2025
In today’s data-driven IT landscape, managing ML lifecycles and operations is converging. On this podcast, we’ll explore how end-to-end ML lifecycle practices extend to proactive, automation-driven IT operations. We'll discuss key MLOps concepts—CI/CD pipelines, feature stores, model monitoring—and how they power anomaly detection, event correlation, and automated remediation.
# AI Managed Services
# Data Analytics
# PwC
Kopal Garg
Kopal Garg · Nov 12th, 2025
LLMs can perform complex tasks like drafting contracts or answering medical questions, but without safeguards, they pose serious risks—like leaking PII, giving unauthorized advice, or enabling fraud. NVIDIA’s NeMo Guardrails provides a modular safety framework that enforces AI safety through configurable input and output guardrails, covering risks such as PII exposure, jailbreaks, legal liability, and regulatory violations. In high-stakes areas like healthcare, it blocks unauthorized diagnoses and ensures HIPAA/FDA compliance. Each blocked action includes explainable metadata for auditing and transparency, turning AI safety from a black-box filter into configurable, measurable infrastructure.
# LLMs
# NeMo Guardrails
# PII
# HIPAA/FDA
Andy Pernsteiner
Demetrios Brinkmann
Andy Pernsteiner & Demetrios Brinkmann · Nov 11th, 2025
Most AI projects don’t fail because of bad models; they fail because of bad data plumbing. Andy Pernsteiner joins the podcast to talk about what it actually takes to build production-grade AI systems that aren’t held together by brittle ETL scripts and data copies. He unpacks why unifying data - rather than moving it - is key to real-time, secure inference, and how event-driven, Kubernetes-native pipelines are reshaping the way developers build AI applications. It’s a conversation about cutting out the complexity, keeping data live, and building systems smart enough to keep up with your models.
# GPU Clusters
# Production-grade AI Systems
# VAST Data
Siddharth Bidasaria
Demetrios Brinkmann
Siddharth Bidasaria & Demetrios Brinkmann · Nov 5th, 2025
Demetrios Brinkmann talks with Siddharth Bidasaria about Anthropic’s Claude code — how it was built, key features like file tools and Spotify control, and the team’s lean, user-focused approach. They explore testing, subagents, and the future of agentic coding, plus how users are pushing its limits.
# Claude Code
# Agentic Coding
# Anthropic
Deploying AI agents in enterprises is complex, balancing security, scalability, and usability. This post compares deployment paths on Google Cloud—highlighting Cloud Run with IAP as the most secure and flexible option—and shows how teams can build powerful agents with ADK without losing the human touch.
# AI Agent
# Agentops
# Generative AI Tools
# Data Science
# Artificial Intelligence
Code of Conduct