Maximum Likelihood Estimation Conceptual Questions

Metronome

Junior Member
Joined
Jun 12, 2018
Messages
154
Suppose I am told the outcomes (x1x_1 through xnx_n) of nn rolls (X1X_1 through XnX_n) of a fair, θ\theta-sided die, and want to find a point estimate of θ\theta.

Let's state the problem without using the lexeme "likely" in order to eliminate one layer of confusion, since its technical and colloquial meanings differ. It seems to me that the goal here is to choose the possible value of θ\theta which has the greatest probability of being its actual value, given the information I have about the rolls, and that this goal translates into the mathematical notation maxθPr(θ(i=1nXi=xi))\max_\theta \Pr(\theta | (\bigcap_{i = 1}^n X_i = x_i)) However, the prescription of the maximum likelihood principle, as I understand it, is to choose the possible value of θ\theta which makes the information I have about the rolls most probable, and that this prescription translates into the mathematical notation maxθPr((i=1nXi=xi)θ)\max_\theta \Pr((\bigcap_{i = 1}^n X_i = x_i) | \theta)
1) Is the point of the maximum likelihood principle that both of these maximizations produce the same answer? If not, then why is the latter the correct one to solve, even while the former seems to better capture the English description of the problem statement?

2) As explained here, the correct answer is that the point estimate of θ\theta is the greatest xix_i observed among the rolls. However, this does not make intuitive sense for small values of nn. For example, after a single roll is observed, say it's x1=15x_1 = 15, it seems more reasonable to guess that the die has 3030 sides than 1515. Is this correct? Does MLE require a suitable sample size to arrive at the correct answer? Is there a more general way to solve this problem for any sample size that agrees with MLE when its sample size requirements are met?
 
Top