Home
MLOps Community
The MLOps Community is where machine learning practitioners come together to define and implement MLOps.
Our global community is the default hub for MLOps practitioners to meet other MLOps industry professionals, share their real-world experience and challenges, learn skills and best practices, and collaborate on projects and employment opportunities. We are the world's largest community dedicated to addressing the unique technical and operational challenges of production machine learning systems.
Events
4:00 PM - 5:00 PM, Apr 10 GMT
Coding Agents Lunch & Learn: Skill Building Workshop (From Idea to Evaluation)



4:00 PM - 5:00 PM, Apr 3 GMT
Coding Agents Lunch & Learn, Session 7
4:00 PM - 6:45 PM, Mar 26 GMT
Ship Agents: A Virtual Conference
4:30 PM - 5:30 PM, Mar 25 GMT
Operationalizing AI Agents: From Experimentation to Production
Content
Video
Most people cripple coding agents by micromanaging them—reviewing every step and becoming the bottleneck.
The shift isn’t to better supervise agents, but to design systems where they work well on their own: parallelized, self-validating, and guided by strong processes.
Done right, you don’t lose control—you gain leverage. Like paving roads for cars, the real unlock is reshaping the environment so AI can move fast.
Apr 7th, 2026 | Views 38
Blog
This blog walks through how to build an AI agent that can meaningfully use a unified collection of multimodal assets like speaker bios, talk titles or descriptions, and eventually their PDF or video content, by pairing the right tools with the right memory design. It demonstrates how retrieval, parsing, and reasoning components must be engineered so the agent can navigate relationships, interpret metadata, and answer higher‑order questions with accuracy. By grounding the workflow in a multimodal database like ApertureDB, the agent gains reliable access to structured context, enabling richer insights across any real‑world content collection.
Apr 7th, 2026 | Views 26
Video
Kashish zooms out to discuss a universal industry pattern: how infrastructure—specifically data loading—is almost always the hidden constraint for ML scaling.
The conversation dives deep into a recent architectural war story. Kashish walks through the full-stack profiling and detective work required to solve a massive GPU starvation bottleneck. By redesigning the Petastorm caching layer to bypass CPU transformation walls and uncovering hidden distributed race conditions, his team boosted GPU utilization to 60%+ and cut training time by 80%. Kashish also shares his philosophy on the fundamental trade-offs between latency and efficiency in GPU serving.
Apr 3rd, 2026 | Views 28

