Sign In Start Free Trial
Account

Add to playlist

Create a Playlist

Modal Close icon
You need to login to use this feature.
  • Debugging Machine Learning Models with Python
  • Toc
  • feedback
Debugging Machine Learning Models with Python

Debugging Machine Learning Models with Python

By : Ali Madani
4.9 (16)
close
Debugging Machine Learning Models with Python

Debugging Machine Learning Models with Python

4.9 (16)
By: Ali Madani

Overview of this book

Debugging Machine Learning Models with Python is a comprehensive guide that navigates you through the entire spectrum of mastering machine learning, from foundational concepts to advanced techniques. It goes beyond the basics to arm you with the expertise essential for building reliable, high-performance models for industrial applications. Whether you're a data scientist, analyst, machine learning engineer, or Python developer, this book will empower you to design modular systems for data preparation, accurately train and test models, and seamlessly integrate them into larger technologies. By bridging the gap between theory and practice, you'll learn how to evaluate model performance, identify and address issues, and harness recent advancements in deep learning and generative modeling using PyTorch and scikit-learn. Your journey to developing high quality models in practice will also encompass causal and human-in-the-loop modeling and machine learning explainability. With hands-on examples and clear explanations, you'll develop the skills to deliver impactful solutions across domains such as healthcare, finance, and e-commerce.
Table of Contents (26 chapters)
close
1
Part 1:Debugging for Machine Learning Modeling
5
Part 2:Improving Machine Learning Models
10
Part 3:Low-Bug Machine Learning Development and Deployment
15
Part 4:Deep Learning Modeling
19
Part 5:Advanced Topics in Model Debugging

Transformers for language modeling

Transformers were introduced in a famous paper called Attention is All You Need (Vaswani et al., 2017) as a new approach for sequence-to-sequence data modeling tasks such as translating statements from one language into another (that is, machine translation). These models are built on top of the idea of self-attention, which helps the model pay attention to other important parts of a sentence or sequence of information in the learning process during training. This attention mechanism helps the models better understand the relationships between the elements of input sequences – for example, between the words in the input sequences in language modeling. Models built using transformers usually work better than ones built using predecessor techniques such as Long Short Term Memory (LSTM) and Recurrent Neural Networks (RNNs) (Vaswani et al., 2017; Devlin et al., 2018).

Figure 13.6 shows four traditional problems in language modeling that have...

bookmark search playlist download font-size

Change the font size

margin-width

Change margin width

day-mode

Change background colour

Close icon Search
Country selected

Close icon Your notes and bookmarks

Delete Bookmark

Modal Close icon
Are you sure you want to delete it?
Cancel
Yes, Delete