A logical and sequential roadmap to understanding the advanced concepts in training deep neural networks.
We will break our discussion into 4 logical parts that build upon each other. For the best reading experience, please go through them sequentially:
1. What is Vanishing Gradient? Why is it a problem? Why does it happen?
2. What is Batch Normalization? How does it help in Vanishing Gradient?
3. How does ReLU help in Vanishing Gradient?
4. Batch Normalization for Internal Covariate Shift
First, let’s understand what vanishing means:
Vanishing means that it goes towards 0 but will never really be 0.
Learn to correctly interpret the coefficients of Logistic Regression and in the process naturally derive its cost function — the Log Loss!
Models like Logistic Regression often win over their complex counterpart models when explainability and interpretability are crucial to the solution. Despite this, unfortunately, Logistic Regression coefficients are not so easy to interpret as the usual Linear Regression coefficients.
Imagine choosing Logistic Regression for sole reasons of explainability yet presenting wrong descriptions to the business stakeholders. Ouch, not a pleasant scenario definitely!
In this blog, I have described how we can derive the interpretation of logistic regression coefficients naturally…
Ever felt curious about this well-known axiom: “Always scale your features”? Well, read on to get a quick graphical and intuitive explanation!
I am sure all of us have seen this popular axiom in machine learning: Always scale your features before training!
While most of us know its practical importance, not many of us are aware of the underlying mathematical reasons.
In this super short blog, I have explained what happens behind the scene with our favorite Gradient Descent algorithm when it is fed with features having very different magnitudes.
Let’s say we are trying to predict the life expectancy…
Do you mean the cell that contains the "center" of an object?
…e current cell. Unlike YOLO v2, we will use binary cross-entropy instead of mean square error here. In the ground truth, objectness is always 1 for the cell that contains an object, and 0 for the cell that doesn’t contain any object. By measuring this obj_loss, we can gradually teach the network to detect a region of interest. In t…
Ethan Yanjia Li
Ace your ML interview by quickly understanding which real-world use cases demand higher precision, and which ones demand a higher recall and why?
All machine learning interviews expect you to understand the practical application of precision-recall tradeoff in real-world use cases, beyond just the definitions and formulas.
I have tried to capture this essence by defining a 🔑 “secret key” that you can exploit to ace your next ML interview and impress your interviewer by providing articulate justifications!
💡 Precision measures that out of all the positive predicted examples, how many detections were correct?
💡 Recall measures that out of…
Focal Loss explained in simple words to understand what it is, why is it required and how is it useful — in both an intuitive and mathematical formulation.
Most object detector models use the Cross-Entropy Loss function for their learning. The idea is to have a loss function that predicts a high probability for a positive example, and a low probability for a negative example, so that using a standard threshold, of say 0.5, we can easily differentiate between the two classes. …
Empower your deep learning models by harnessing some immensely powerful image processing algorithms.
Many deep learning courses start with an introduction to the basic image processing techniques (like resizing, cropping, color-to-grayscale, rotation, etc.) but only provide a cursory glance at these concepts.
In this journey, we often discount the importance of some immensely powerful image processing algorithms which can do wonders to our final model predictions.
For folks who are familiar with the general ML paradigm, you know how vital data cleaning and feature engineering is for the success of any model. …
In the model training phase, a model learns its parameters. But there are also some secret knobs, called hyperparameters, that the model cannot learn on its own — these are left to us to tune. Tuning hyperparameters can significantly improve model performance. Unfortunately, there is no definite procedure to calculate these hyperparameter values. This is why hyperparameter tuning is often regarded as an art than science.
In this article, I discuss the 3 most popular hyperparameter tuning algorithms — Grid search, Random search, and Bayesian optimization.
Model training is a process through which a model learns its parameters. Besides this…
With so many rampant advances taking place in Natural Language Processing (NLP), it can sometimes become overwhelming to be able to objectively understand the differences between the different models.
It is important to understand not only how these models differ from each other, but also how one model overcomes the shortcomings of another.
Below I have drawn out a comparison between two very popular models — Word2Vec and BERT.
Word2Vec models generate embeddings that are context-independent: ie - there is just one vector (numeric) representation for each word. …
AWS ML Specialist | Instructor & Mentor for Data Science