Episode 6: What in the world is Kullback-Leibler Divergence?

This week I discovered a new blog on topics related to Bayesian Theory and this is everything I would like to remember from a post related to Kullback-Leibler Divergence.

Concepts:

If we have a population of subjects, and they can have any number of diseases from 0 to 10, as shown in the figure above, then n = 10. We don’t know the probability p for our binomial distribution, so we calculate the mean of our observed data Using E[X] = np, we calculate p which comes out to be 4.7/10 = 0.47. So binomial distribution can be constructed using:


Where k goes from 0 to 10.

Source: Medium, A new Tool to your Toolkit, KL Divergence at Work

P is actual distribution and q is approximated distribution. Since expectation E[x] is defined as the product of x and its probability, K-L divergence is the expectation of the difference between log of actual distribution and log of approximated distribution.

Thought of the Week:

This week I found myself in a rut when I sat down to organize my Machine Learning notes. Taking a course in Machine Learning, I have come to realize, is like going to Ikea. You like everything and you want to buy everything but you rarely need all those things at the same time. So, earlier this week, I rolled up my sleeves, picked up my data-mining shovel and decided to dig deep into the topic of bias-variance tradeoff, only to end up having a vertigo as I jumped from one concept to another, with a hundred different tabs opened up in my browser. But one good thing that came out of it was discovering a new blog on probability theory called Count Bayesie” (which is already up on my favorite list of blogs), when I was trying to find out how to expand the expectation of mean squared error (geeky segue alert!). Discovering a good resource in data science makes me realize how far I am from where I want to be in this field, but it is also like discovering a new trail along the way, a detour which might make this long, arduous journey plagued with frequent fits of self-scrutiny and imposter syndrome, more fun and exploratory. There is so much to learn and you get a day everyday to do it. What more could one ask for? Happy reading!

See you next week!

References:

[1] Kullback-Leibler Divergence Explained
[2] A new divergence measure for basic probability assignment and its applications in extremely uncertain environments
[3] The AIC Criterion and Symmetrizing the Kullback–Leibler Divergence
[4] A New Kullback–Leibler VAD for Speech Recognition in Noise
[5] A Kullback-Leibler Divergence Based Kernel for SVM Classification in Multimedia Applications

Share this: