httpwwwamstatorgpublicationsjsev12n2zhupdfCopyrightc2004byMuZhuandArthurYLuallrightsreservedThistextmaybefreelysharedamongindvidualsbutitmaynotberepublishedinanymediumwithoutexpresswrittenconsen ID: 885450
Download Pdf The PPT/PDF document "TheCounterintuitiveNoninformativePriorfo..." is the property of its rightful owner. Permission is granted to download and print the materials on this web site for personal, non-commercial use only, and to display it on your personal computer provided you do not modify the materials and that you retain all copyright notices contained in the materials. By downloading content from our website, you accept the terms of this agreement.
1 TheCounter-intuitiveNon-informativePrior
TheCounter-intuitiveNon-informativePriorfortheBernoulliFamilyMuZhuUniversityofWaterlooArthurY.LuRenaissanceTechnologiesCorp.JournalofStatisticsEducationVolume12,Number2(2004), http://www.amstat.org/publications/jse/v12n2/zhu.pdf Copyrightc 2004byMuZhuandArthurY.Lu,allrightsreserved.Thistextmaybefreelysharedamongindviduals,butitmaynotberepublishedinanymediumwithoutexpresswrittenconsentfromtheauthorandadvancenoticationoftheeditor. KeyWords:BetaDistribution;Conjugatepriors;Maximumlikelihoodestimation;Poste-riormean.AbstractInBayesianstatistics,thechoiceofthepriordistributionisoftencontroversial.Dierentrulesforselectingpriorshavebeensuggestedintheliterature,which,sometimes,producepriorsthataredicultforthestudentstounderstandintuitively.Inthisarticle,weuseasimpleheuristictoillustratetothestudentstherathercounter-intuitivefactthat atpriorsarenotnecessarilynon-informative;andnon-informativepriorsarenotnecessarily at.1IntroductionInBayesiananalysis,selectingpriorsusingJereys'rule(e.g.,Tanner1996;Gelman,etal.1995)canyieldsomerathercounter-intuitiveresultsthatarehardforstudentstograsp.Considerthefollowingsimplescenario:LetX1;X2;:::;Xnbei.i.d.observationsfromtheBernoulli(p)distribution.Toestimatep,aBayesiananalystwouldputapriordistributiononpandusetheposteriordistributionofptodrawvariousconclusions,e.g.,estimatingpwiththeposteriormean.Whenthereisnostrongprioropiniononwhatpis,itisdesirabletopickapriorthatisnon-informative. 1 Inthissimpl
2 ecase,itismostintuitivetousetheuniformdi
ecase,itismostintuitivetousetheuniformdistributionon[0,1]asanon-informativeprior;itisnon-informativebecauseitsaysthatallpossiblevaluesofpareequallylikelyapriori.However,anon-informativepriorconstructedusingJerey'sruleisoftheform(seee.g.,Gelman1995)(p)/1 p p(1 p): (1) Jerey'sruleismotivatedbyaninvarianceargument:Supposeonepicksp(p)asthepriorforpaccordingtoacertainrule.Inorderforp(p)tobenon-informative,itisarguedthattheparameterizationmustnotin uencethechoiceofp(p),i.e.,ifonere-parameterizestheproblemintermsof=h(p),thentherulemustpick()=p(h 1())dp dasthepriorfor.Givenp,letf(xjp)bethelikelihoodfunction.Jerey'sruleistopickp(p)/(I(p))1 2asaprior,whereI(p)= Ed2logf(xjp) dp2istheFisherInformation.Toseethatthisisinvariantwithrespecttoparameterization,supposewere-parameterizeintermsof=h(p),thenI()= Ed2logf(xj) d2= E d2logf(xjp) dp2dp d2!=I(p)dp d2:ApplyingJerey'srule,onewouldpickaprioronas()/(I())1 2=(I(p))1 2dp d=p(p)dp d;whichsatisiestheinvarianceargument.Jerey'spriorforthissimpleproblemcanbequitecounter-intuitive.Undertheprior(1),itappearsthatsomevaluesofparemorelikelythanothers(seee.g.,Figure1).Therefore 2 intuitively,itappearsthatthispriorisactuallyquiteinformative.Thisisaverydicultpointtoexplaintothestudents.Inthisarticle,weconstructasimple(albeitnaive)argumentandillustratetothestudentswhytheuniformpriorisnotnecessarilythemostnon-i
3 nformative.WewillnotrelyontheFisherinfor
nformative.WewillnotrelyontheFisherinformationorJereys'invarianceargument.Instead,werelyonaverysimpleandnaiveheuristictojudgethenon-informativenessofaprior:Sincethemaximumlikelihoodestimator(MLE)isnotaectedbyanyprioropinion,wesimplyask:isthereapriorwhichwouldproduceaBayesianestimate(e.g.,posteriormean)thatcoincideswiththeMLE?Ifso,thatpriorcouldberegardedasnon-informativesincetheprioropinionexertsnoin uenceonthenalestimatewhatsoever.Usingthisnaiveheuristic,wecanseethattheuniformpriorisactuallymoreinformativethanJereys'prior(1);whereastheleastinformativeprioris,surprisinglyenough,anextremely\opinionated"distributionapproachingtwopointmassesat0and1!Weemphasizeherethatitisnotourintentiontoimplythatournaiveheuristicisthebestorevenanappropriatepointofviewforjudgingthenon-informativenessofdierentpriorsinBayesiananalysis.Weonlyprovidethisargumentasaninterestingdemonstrationthatcanbeusedintheclassroom.2TheMaximumLikelihoodEstimator(MLE)Withoutconsideringanyprioropinion,atypicalapproachforestimatingpisthemethodofmaximumlikelihood.LetXibeaBernoullirandomvariablewithP(Xi=1)=p;P(Xi=0)=1 p;thelog-likelihoodfunctionfortheBernoullidistributionisl(p)=nXi=1logpxi(1 p)1 xi=nXi=1xilogp+(1 xi)log(1 p):Weshallwritex=Pni=1xithroughoutthearticle.Tomaximizethislog-likelihood,therstorderconditionisx p n x 1 p=0;whichgives^pmle=x n:3TheBayesianEstimatorTheBayesianapproachtoestimationstartswithapriordistributionontheparameterofinterest.Often,weha
4 venopriorknowledgeonp.Tore ectsuchlackof
venopriorknowledgeonp.Tore ectsuchlackofknowledge,themostintuitivechoiceistoputauniformprioronp,i.e.,0(p)=1forp2[0;1].Thissaysthat,apriori,pcouldbeanythingbetween0and1withequalchance.Then,theposterior 3 distributionofpisgivenby(Bayes'Theorem):1(pjx1;:::;xn)=f(x1;:::;xnjp)0(p) R10f(x1;:::;xnjp)0(p)dp:Weshallndthisposteriordistributionmoregenerallybelowusingtheideaoftheconjugateprior.3.1TheBetaConjugatePriorConsidertheBeta(;)distributionasthepriorforp,i.e.,0(p)= (+) () ()p 1(1 p)