Articles → MACHINE LEARNING → Information Gain In Machine Learning
Information Gain In Machine Learning
Purpose
Significance
Formula
- S = the original dataset (before the split)
- A = the attribute for which IG is being calculated
- v∈Values(A)=each possible value of attribute A
- Sv = the subset of 𝑆 where attribute 𝐴 has value
- ∣Sv∣ = number of samples in subset Sv
- ∣S∣=total number of samples
- Entropy(S) = measure of impurity in dataset 𝑆
Sample Dataset
ID | Outlook | Temperature | Humidity | Windy | Play Tennis |
---|
1 | Sunny | Hot | High | False | No |
2 | Sunny | Hot | High | True | No |
3 | Overcast | Hot | High | False | Yes |
4 | Rain | Mild | High | False | No |
5 | Rain | Cool | Normal | False | No |
6 | Rain | Cool | Normal | True | No |
7 | Overcast | Cool | Normal | True | Yes |
8 | Sunny | Mild | High | False | Yes |
Entropy Of Full Dataset
Entropy Of Sunny
Entropy Of Overcast
Entropy Of Rain
Weighted Average Entropy For Outlook
Information Gain (Outlook)
Posted By - | Karan Gupta |
|
Posted On - | Tuesday, September 2, 2025 |