Creating an ML model is just a starting point. To bring the technology into production service, you need to solve various real-world issues such as building a data pipeline for continuous training, automated validation of the model, version control of the model, scalable serving infra, and ongoing operation of the ML infra with monitoring and alerting. Kaz Sato discusses ML ops concepts and explains how to combine Kubeflow pipelines to build a production ML service infra.
Kaz Sato is a staff developer advocate on the cloud platform team at Google, where he leads the developer advocacy team for machine learning and data analytics products such as TensorFlow, the Vision API, and BigQuery. Kaz has been leading and supporting developer communities for Google Cloud for over seven years. He’s a frequent speaker at conferences, including Google I/O 2016, Hadoop Summit 2016 San Jose, Strata and Hadoop World 2016, and Google Next 2015 NYC and Tel Aviv, and he has hosted FPGA meetups since 2013.
©2019, O'Reilly Media, Inc. • (800) 889-8969 or (707) 827-7019 • Monday-Friday 7:30am-5pm PT • All trademarks and registered trademarks appearing on oreilly.com are the property of their respective owners. • email@example.com