MLOps Community
+00:00 GMT
Sign in or Join the community to continue

PyTorch: Bridging AI Research and Production

Posted Nov 15, 2021 | Views 301
# PyTorch
# PyTorch Ecosystem
# AI Models
# fireworks.ai
Share
speakers
avatar
Dmytro Dzhulgakov
CTO, Co-Founder @ Fireworks AI

Dmytro (Dima) Dzhulgakov is the co-founder and CTO of Fireworks.ai, which focuses on the transition to AI-powered business via interactive experimentation and a production platform centered around PyTorch technologies. Fireworks.ai offers high-performance low-cost LLM inference service that helps to try out and productionize large models.

Dmytro is one of PyTorch core maintainers. Previously he helped to bring PyTorch from a research framework to numerous production applications across Meta's AI use cases and broader industry.

+ Read More
avatar
Demetrios Brinkmann
Chief Happiness Engineer @ MLOps Community

At the moment Demetrios is immersing himself in Machine Learning by interviewing experts from around the world in the weekly MLOps.community meetups. Demetrios is constantly learning and engaging in new activities to get uncomfortable and learn from his mistakes. He tries to bring creativity into every aspect of his life, whether that be analyzing the best paths forward, overcoming obstacles, or building lego houses with his daughter.

+ Read More
avatar
Vishnu Rachakonda
Data Scientist @ Firsthand

Vishnu Rachakonda is the operations lead for the MLOps Community and co-hosts the MLOps Coffee Sessions podcast. He is a machine learning engineer at Tesseract Health, a 4Catalyzer company focused on retinal imaging. In this role, he builds machine learning models for clinical workflow augmentation and diagnostics in on-device and cloud use cases. Since studying bioengineering at Penn, Vishnu has been actively working in the fields of computational biomedicine and MLOps. In his spare time, Vishnu enjoys suspending all logic to watch Indian action movies, playing chess, and writing.

+ Read More
SUMMARY

Over the past few years, PyTorch became the tool of choice for many AI developers ranging from academia to industry. With the fast evolution of state-of-the-art in many AI domains, the key desired property of the software toolchain is to enable the swift transition of the latest research advances to practical applications. In this coffee session, Dmytro discusses some of the design principles that contributed to this popularity, how PyTorch navigates inherent tension between research and production requirements, and how AI developers can leverage PyTorch and PyTorch ecosystem projects for bringing AI models to their domain.  

+ Read More
TRANSCRIPT

Quotes

If you make a lot of people more productive, that accelerates progress overall.

Whatever dimension you look at starting from hardware and capabilities to scale of a number of developers, applications, and different domains which within Facebook obviously within the industry gets applied.

As I mentioned, it's a community project, it's not on the Facebook Meta project. I feel that if you work with the community, you can collaborate very actively with people from academia, industry, and different hardware vendors. It feels pretty exciting and very different almost like a running mini startup project.

The MLOps space grew so much. There's a lot of exciting opportunities and a lot of problems to solve.

Always operate in product mode. When people are pushing the boundary, people can unlock their ideas, unblock themselves, try to create new trends, and fundamentally push efforts forward so definitely there's this mindset of product building mindset.

It's really important to make a modeler happy. This is what fundamentally drives the innovation of AI so far and this trend hasn't been deaccelerating so far.

Basically, if you look at the state of models over the years, there is a lot of incremental improvements, a number of papers published absolutely crazy, all the benchmarks have been pushed but also new revolution has been coming.

AI is basically an Applied Science field. The only way to progress is to increase the federation and try new ideas. If you try new ideas faster, you end up with more progress on average.

In the early days, pushing the modeler first was really crucial because that's where the innovation space happens to flow from research to production.

As some techniques become more mature and widely applicable then it makes sense to upstream and absorb them into play PyTorch package itself. They can be more easily accessible to a broader audience.

ML space is still very early which means that it's very fragmented or partitioned, so making integration easier is actually often the highest leverage thing to removing friction from the other side. That's one big theme which in general the Pytorch community is trying to enable.

It's easy to get something working and then making it fast incrementally afterward. I think while still staying in pretty much the same environment, that when this kind of pivotal PyTorch as a framework serving both instead of trying to export.

War stories are like the constant flow of new ideas, models, and pipelines between 2 worlds and you should not think of it as a mountain to another mountain to build different solutions.

People are trying to use so many different hardware platforms that are popping up and it's also like a second order of that which is how to enable different hardware vendors to build solutions where they can do this optimization whether you have a creative or latest one.

I would imagine that more and more people are doing ML as a part of their presence, especially in products that have benefited from this that expands quite a bit so more and more regular engineers should learn this space.

Our job is to make the best tools to enable this innovation process both for folks that are pretty far trying to advance ideas pushing MLOps to be flexible and also for people who are just getting into this data-driven mindset by providing a higher-level abstraction.

+ Read More

Watch More

From Research to Production: Fine-Tuning & Aligning LLMs // Philipp Schmid // AI in Production
Posted Feb 25, 2024 | Views 1.3K
# LLM
# Fine-tuning LLMs
# dpo
# Evaluation
Generative AI Agents in Production: Best Practices and Lessons Learned // Patrick Marlow // Agents in Production
Posted Nov 15, 2024 | Views 2.3K
# Generative AI Agents
# Vertex Applied AI
# Agents in Production