Multimodal models are a game changer, but do you actually understand how they work? What about embeddings and RAG? Check out our roundup of the top AI and MLOps articles for November 2023!
Machine Learning Engineering Guides and Tools
Stas Bekman has extensive experience training LLMs from his time at Hugging Face. His repo of ML engineering guides and tools can help you get started on your LLM journey: https://github.com/stas00/ml-engineering
Multimodality and Large Multimodal Models (LMMs)
Multimodal models are step change in AI. Chip Huyen explains the fundamentals of a multimodal system and current research advancing LMMs: https://huyenchip.com/2023/10/10/multimodal.html
What is Retrieval-Augmented Generation (RAG)?
Retrieval-augmented generation is all the rage, but do you actually understand RAG architecture and its benefits? Valeriia Kuka breaks it down for the rest of us: https://www.turingpost.com/p/rag
How to Monitor LLMOps Performance with Drift
Like predictive ML models, LLM performance also degrades over time. Learn the two types of performance problems and how to identify drift in prompts and responses: https://www.fiddler.ai/blog/how-to-monitor-llmops-performance-with-drift
State of AI Report 2023
AI is much more than LLMs. From weather prediction to self-driving cars to music generation, the amazing State of AI report covers it all — and yes, that includes a whole lot of LLM research too: https://www.stateof.ai/
How LinkedIn Is Using Embeddings to Up Its Match Game for Job Seekers
LinkedIn uses embedding-based retrieval to power search and recommendations across the platform, from the newsfeed to notifications to job openings. Their ML team shares how they built infrastructure to incorporate embeddings at scale: https://engineering.linkedin.com/blog/2023/how-linkedin-is-using-embeddings-to-up-its-match-game-for-job-se
Embeddings: What they are and why they matter
Embeddings. You've probably heard of them. You may have worked with them. But how do they work and why do they matter? https://simonwillison.net/2023/Oct/23/embeddings/
Retrieval Augmented Generation at scale — Building a distributed system for synchronizing and ingesting billions of text embeddings
Retrieval-augmented generation (RAG) is relatively straight forward, but the difficulty lies in scaling and becoming production-ready. And what if you're trying to ingest a billion records? Here's a great guide on how to get it done: https://medium.com/@neum_ai/retrieval-augmented-generation-at-scale-building-a-distributed-system-for-synchronizing-and-eaa29162521
Stay up to date on everything AI and MLOps by signing up for our newsletter below.