
Your one-stop shop for all Changelog podcasts. Weekly shows about software development, developer culture, open source, building startups, artificial intelligence, shipping code to production, and the people involved. Yes, we focus on the people. Everything else is an implementation detail.
Testing ML systems (Practical AI #74)
Production ML systems include more than just the model. In these complicated systems, how do you ensure quality over time, especially when you are constantly updating your infrastructure, data and models? Tania Allard joins us to discuss the ins and outs of testing ML systems. Among other things, she presents a simple formula that helps you score your progress towards a robust system and identify problem areas.
Changelog++ members support our work, get closer to the metal, and make the ads disappear. Join today!
Featuring
- Tania Allard – Twitter, GitHub, Website
- Chris Benson – Twitter, GitHub, LinkedIn, Website
- Daniel Whitenack – Twitter, GitHub, Website
Notes and Links
- “What’s your ML score” talk
- “Jupyter Notebooks: Friends or Foes?” talk
- Joel Grus’s episode: “AI code that facilitates good science”
- Papermill
- nbdev
- nbval
- DevOps for dummies
Something missing or broken? PRs welcome!