# Bayesian

Least square method: deciding model parameters by minimizing RSS (residual sum of squares)
ML (maximum likelihood) estimation: estimating model parameters by finding the parameter values that maximize the likelihood of making the observations given the parameters
MAP (maximum a posteriori) estimation

## Bayes’ Theorem where A and B are events and P(B) ≠ 0. : conditional probability of event A given that B is true.  : probability of event A, probability of event B without regarding each other. : conditional probability of event A given that B is true.   ### example

Suppose that a probability of having disease A is 0.5%.
And suppose that a test is 99% sensitive (true positive rate), 95% specific (true negative rate).
If you’re detected as positive by the test. What is the probability that you have disease A? Even if you’re diagnosed as positive, the probability of having disease is only around 9%

### Bayesian inference : parameters of probability distribution : observed data (fixed) : posterior probability : likelihood : prior probability : marginal likelihood or normalization constant Try to find which maximizes http://mlss.tuebingen.mpg.de/2015/slides/ghahramani/lect1bayes.pdf

http://hosho.ees.hokudai.ac.jp/~kubo/stat/2010/Qdai/b/kuboQ2010b.pdf