Monitoring and explainability of models in production
2020-07-13Code Available2· sign in to hype
Janis Klaise, Arnaud Van Looveren, Clive Cox, Giovanni Vacanti, Alexandru Coca
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/SeldonIO/alibi-detectOfficialIn papertf★ 2,505
Abstract
The machine learning lifecycle extends beyond the deployment stage. Monitoring deployed models is crucial for continued provision of high quality machine learning enabled services. Key areas include model performance and data monitoring, detecting outliers and data drift using statistical techniques, and providing explanations of historic predictions. We discuss the challenges to successful implementation of solutions in each of these areas with some recent examples of production ready solutions using open source tools.