Articles → MACHINE LEARNING → Bayes Theorem In Machine Learning
Bayes Theorem In Machine Learning
Probability
- 0 = impossible event
- 1 = certain event
Formula Of Probability
- Probability of getting a 3 = 1/6
- Probability of getting an even number = 3/6=0.5
Independent And Dependent Events
Joint Probability
Union
A: Even number = {2, 4, 6}
B: Multiple of 3 = {3, 6}
P(A) = 3/6 = 1/2
P(B) = 2/6 = 1/3
P(A∩B) = (1/2 X 1/3) = 1/6
Union = 1/2 + 1/3 – 1/6 = 2/3 = 66.7%
Conditional Probability
P(B) = 40/100 = 0.4
P(A∩B) = 10/100 = 0.1
P(A|B) = 0.1/0.4 = 0.25
Bayer’S Theorem
Parameter | Description |
---|
P(A) | Probability of an event before seeing evidence (called prior probability). |
P(B) | New evidence |
P(B|A) | The likelihood: how probable it is to observe the evidence B if A were true. |
P(A|B) | Updated probability after observing the evidence B (called Posterior probability). |
Posted By - | Karan Gupta |
|
Posted On - | Thursday, September 25, 2025 |