Share
Machine Learning Design Patterns: Solutions to Common Challenges in Data Preparation, Model Building, and Mlops
Valliappa Lakshmanan
(Author)
·
Sara Robinson
(Author)
·
Michael Munn
(Author)
·
O'Reilly Media
· Paperback
Machine Learning Design Patterns: Solutions to Common Challenges in Data Preparation, Model Building, and Mlops - Lakshmanan, Valliappa ; Robinson, Sara ; Munn, Michael
£ 47.69
£ 52.99
You save: £ 5.30
Choose the list to add your product or create one New List
✓ Product added successfully to the Wishlist.
Go to My WishlistsIt will be shipped from our warehouse between
Wednesday, May 08 and
Thursday, May 09.
You will receive it anywhere in United Kingdom between 1 and 3 business days after shipment.
Synopsis "Machine Learning Design Patterns: Solutions to Common Challenges in Data Preparation, Model Building, and Mlops"
The design patterns in this book capture best practices and solutions to recurring problems in machine learning. The authors, three Google engineers, catalog proven methods to help data scientists tackle common problems throughout the ML process. These design patterns codify the experience of hundreds of experts into straightforward, approachable advice. In this book, you will find detailed explanations of 30 patterns for data and problem representation, operationalization, repeatability, reproducibility, flexibility, explainability, and fairness. Each pattern includes a description of the problem, a variety of potential solutions, and recommendations for choosing the best technique for your situation. You'll learn how to: Identify and mitigate common challenges when training, evaluating, and deploying ML models Represent data for different ML model types, including embeddings, feature crosses, and more Choose the right model type for specific problems Build a robust training loop that uses checkpoints, distribution strategy, and hyperparameter tuning Deploy scalable ML systems that you can retrain and update to reflect new data Interpret model predictions for stakeholders and ensure models are treating users fairly