Articles → MACHINE LEARNING → Information Gain In Machine Learning
Information Gain In Machine Learning
Purpose
Significance
Formula
- S = the original dataset (before the split)
- A = the attribute for which IG is being calculated
- v∈Values(A)=each possible value of attribute A
- Sv = the subset of 𝑆 where attribute 𝐴 has value
- ∣Sv∣ = number of samples in subset Sv
- ∣S∣=total number of samples
- Entropy(S) = measure of impurity in dataset 𝑆
Sample Dataset
| ID | Outlook | Temperature | Humidity | Windy | Play Tennis |
|---|
| 1 | Sunny | Hot | High | False | No |
| 2 | Sunny | Hot | High | True | No |
| 3 | Overcast | Hot | High | False | Yes |
| 4 | Rain | Mild | High | False | No |
| 5 | Rain | Cool | Normal | False | No |
| 6 | Rain | Cool | Normal | True | No |
| 7 | Overcast | Cool | Normal | True | Yes |
| 8 | Sunny | Mild | High | False | Yes |
Entropy Of Full Dataset
Entropy Of Sunny
Entropy Of Overcast
Entropy Of Rain
Weighted Average Entropy For Outlook
Information Gain (Outlook)
| Posted By - | Karan Gupta |
| |
| Posted On - | Tuesday, September 2, 2025 |