Yichen Qin Estimation Of Mixture Models & Carey E. Priebe .

2y ago
33 Views
4 Downloads
950.29 KB
16 Pages
Last View : 4d ago
Last Download : 3m ago
Upload by : Sabrina Baez
Transcription

This article was downloaded by: [Johns Hopkins University]On: 23 October 2013, At: 14:19Publisher: Taylor & FrancisInforma Ltd Registered in England and Wales Registered Number: 1072954 Registered office: Mortimer House,37-41 Mortimer Street, London W1T 3JH, UKJournal of the American Statistical AssociationPublication details, including instructions for authors and subscription aximum Lq-Likelihood Estimation via theExpectation-Maximization Algorithm: A RobustEstimation of Mixture ModelsaYichen Qin & Carey E. PriebeaaDepartment of Applied Mathematics and Statistics , Johns Hopkins University, 100Whitehead Hall, 3400 North Charles Street , Baltimore , MD , 21210Accepted author version posted online: 28 Apr 2013.Published online: 27 Sep 2013.To cite this article: Yichen Qin & Carey E. Priebe (2013) Maximum Lq-Likelihood Estimation via the Expectation-MaximizationAlgorithm: A Robust Estimation of Mixture Models, Journal of the American Statistical Association, 108:503, 914-928, DOI:10.1080/01621459.2013.787933To link to this article: SE SCROLL DOWN FOR ARTICLETaylor & Francis makes every effort to ensure the accuracy of all the information (the “Content”) containedin the publications on our platform. However, Taylor & Francis, our agents, and our licensors make norepresentations or warranties whatsoever as to the accuracy, completeness, or suitability for any purpose of theContent. Any opinions and views expressed in this publication are the opinions and views of the authors, andare not the views of or endorsed by Taylor & Francis. The accuracy of the Content should not be relied upon andshould be independently verified with primary sources of information. Taylor and Francis shall not be liable forany losses, actions, claims, proceedings, demands, costs, expenses, damages, and other liabilities whatsoeveror howsoever caused arising directly or indirectly in connection with, in relation to or arising out of the use ofthe Content.This article may be used for research, teaching, and private study purposes. Any substantial or systematicreproduction, redistribution, reselling, loan, sub-licensing, systematic supply, or distribution in anyform to anyone is expressly forbidden. Terms & Conditions of access and use can be found at ns

Maximum Lq -Likelihood Estimation via theExpectation-Maximization Algorithm: A RobustEstimation of Mixture ModelsYichen QIN and Carey E. PRIEBEWe introduce a maximum Lq-likelihood estimation (MLqE) of mixture models using our proposed expectation-maximization (EM) algorithm, namely the EM algorithm with Lq-likelihood (EM-Lq). Properties of the MLqE obtained from the proposed EM-Lq are studiedthrough simulated mixture model data. Compared with the maximum likelihood estimation (MLE), which is obtained from the EM algorithm,the MLqE provides a more robust estimation against outliers for small sample sizes. In particular, we study the performance of the MLqEin the context of the gross error model, where the true model of interest is a mixture of two normal distributions, and the contaminationcomponent is a third normal distribution with a large variance. A numerical comparison between the MLqE and the MLE for this gross errormodel is presented in terms of Kullback–Leibler (KL) distance and relative efficiency.Downloaded by [Johns Hopkins University] at 14:19 23 October 2013KEY WORDS:Gross error model; Robustness.1. INTRODUCTIONMaximum likelihood is among the most commonly used estimation procedures. For mixture models, the maximum likelihood estimation (MLE) via the expectation-maximization (EM)algorithm introduced by Dempster, Laird, and Rubin (1977) isa standard procedure. Recently, Ferrari and Yang (2010) introduced the concept of maximum Lq-likelihood estimation(MLqE), which can yield robust estimation by trading bias forvariance, especially for small or moderate sample sizes. Thisarticle combines the MLqE with the EM algorithm to obtain therobust estimation for mixture models, and studies the performance of this robust estimator.In this article, we propose a new EM algorithm—namely anexpectation-maximization algorithm with Lq-likelihood (EMLq) which addresses MLqE within the EM framework. In theEM-Lq algorithm, we propose a new objective function at eachM-step which plays the role that the complete log-likelihoodplays in the traditional EM algorithm. By doing so, we inheritthe robustness of the MLqE and make it available for mixturemodel estimation.Our study focuses on the performance of the MLqE for estimation in a gross error model f0 (x) (1 )f0 (x) ferr (x),where f0 (x) is what we are interested in estimating and ferr (x)is the measurement error component. For simplicity, we consider the object of interest f0 (x) to be a mixture of two normaldistributions. And ferr (x) is a third normal distribution with alarge variance. We will examine the properties of the MLqE,in comparison to that of the MLE, at different levels of thecontamination ratio .The measurement error problem is one of the most practicalproblems in Statistics. Let us consider that some measurementsX (X1 , X2 , . . . , Xn ) are produced by a scientific experiment.Yichen Qin is PhD student (E-mail: yqin2@jhu.edu) and Carey E. Priebeis Professor (E-mail: cep@jhu.edu), Department of Applied Mathematics andStatistics, Johns Hopkins University, 100 Whitehead Hall, 3400 North CharlesStreet, Baltimore, MD 21210. This work is partially supported by NationalSecurity Science and Engineering Faculty Fellowship (NSSEFF) and JohnsHopkins University Human Language Technology Center of Excellence (JHUHLT COE). The authors thank the associate editor and two referees for insightfulcomments that greatly improved the article.X has a distribution fθ with a interpretable parameter θ thatwe are interested in. However, we do not observe X directly.Instead, we observe X (X1 , X2 , . . . , Xn ) where most of theXi Xi , but there are a few outliers. In other words, X is Xcontaminated with gross errors which are mostly due to eitherhuman error or instrument malfunction. But fθ is still the targetof our estimation (Bickel and Doksum (2007)). To overcomethis problem and still be able to do statistical inference for fθin the mixture model case, we come up with this idea of theEM-Lq.There has been an extensive amount of early work on robustestimation of mixture models and clustering. For example, Peeland McLachlan (2000) used t distributions instead of normaldistributions to incorporate the phenomena of fat tails, and gavea corresponding expectation/conditional maximization (ECM)algorithm which was originally introduced by Meng and Rubin(1993). McLachlan, Ng, and Bean (2006) further formalizedthis idea and applied it to robust cluster analysis. Tadjudin andLandgrebe (2000) made a contribution on robust estimation ofmixture model parameters by using both labeled and unlabeleddata, and assigning different weights to different data points.Garcia-Escudero and Gordaliza (1999) studied the robustnessproperties of the generalized k-means algorithm from the influence function and the breakpoint perspectives. Finally, CuestaAlbertos, Matran, and Mayo-Iscar (2008) applied the trimmedsubsample of the data for fitting mixture models and iterativelyadjusted the subsample after each estimation.The remainder of this article is organized as follows. Section2 gives an introduction to the MLqE along with its advantagescompared to the MLE. Properties of the MLqE for mixturemodels are discussed in Section 3. In Section 4, we present ourEM-Lq, and explain the rationale behind it. The application ofthe EM-Lq in mixture models is introduced and discussed inSection 5. The comparisons of the MLqE (obtained from theEM-Lq) and the MLE based on simulation as well as real data914 2013 American Statistical AssociationJournal of the American Statistical AssociationSeptember 2013, Vol. 108, No. 503, Theory and MethodsDOI: 10.1080/01621459.2013.787933

Qin and Priebe: Maximum Lq -Likelihood Estimation via the Expectation-Maximization Algorithmare presented in Section 6. We address the issue of tuning parameter q in Section 7. We conclude with a discussion and directionsfor future research in Section 8, and relegate the proofs toSection Appendix.2. MAXIMUM Lq -LIKELIHOOD ESTIMATION2.1 Definitions and Basic PropertiesDownloaded by [Johns Hopkins University] at 14:19 23 October 2013First, let us start with the traditional maximum likelihood estimation. Suppose data X follows a distribution with probabilitydensity function fθ parameterized by θ Rd . Given thelikelihood estiobserved data x (x1 , . . . , xn ), the maximum mate is defined as θ̂MLE arg maxθ { ni 1 log f (xi ; θ )}. Similarly, the maximum Lq-likelihood estimate (Ferrari and Yang2010) is defined asθ̂MLqE arg maxθ n Lq (f (xi ; θ )),i 1where Lq (u) (u1 q 1)/(1 q) and q 0. By L’Hopital’srule, when q 1, Lq (u) log(u). The tuning parameter q iscalled the distortion parameter, which governs how distortedLq is away from the log function. Based on this property, weconclude that the MLqE is a generalization of the MLE.Define U (x; θ ) θ log f (x; θ ) fθ (x; θ )/f (x; θ ) and U (x; θ, q) θ Lq (f (x; θ )) U (x; θ )f (x; θ )1 q , we know thatθ̂MLE is a solution of the likelihood equation 0 ni 1 U (xi ; θ ).Similarly, θ̂MLqE is a solution of the Lq-likelihood equation0 n U (xi ; θ, q) i 1n U (xi ; θ )f (xi ; θ )1 q .(1)i 1It is easy to see that θ̂MLqE is a solution to a weighted versionof the likelihood equation that θ̂MLE solves. The weights are proportional to the power transformation of the probability densityfunction, f (xi ; θ )1 q . When q 1, the MLqE puts more weighton the data points with high likelihoods, and less weight on thedata points with low likelihoods. The tuning parameter q adjustshow aggressively the MLqE distorts the weight allocation. TheMLE can be considered as a special case of the MLqE withequal weights.In particular, when f is a normal distribution, our μ̂MLqE andˆ2σ MLqE satisfyμ̂MLqE nn 1i 11σˆ2 MLqE nwii 1 wii 1n wi xi ,(2)wi (xi μ̂MLqE )2 ,(3)915for solving Equations (2) and (3). Details of the algorithm aredescribed in Section 8.2.2 Consistency and Bias-Variance TradeoffBefore discussing the consistency of the MLqE, let us look atthe MLE first. It is well studied that the MLE is quite generally a consistent estimator. Suppose, the true distributionf0 F, where F is a family of distributions; we know thatf0 arg maxg F Ef0 log g(X), which shows the consistency ofthe MLE. However, when we replace the log function with theLq function, we do not have the same property.We first define f (r) , a transformed distribution of f called theescort distribution, asf (r) f (x; θ )r.f (x; θ )r dx(4)We also define F to be a family of distributions that is closedunder such a transformation (i.e., f F, f (r) F). Equippedwith these definitions, we have the following property:(1/q)f0 arg max Ef0 Lq (g(X)).g FThus we see that the maximizer of the expectation of Lqlikelihood is the escort distribution (r 1/q) of the true densityf0 . To also achieve consistency for the MLqE, Ferrari and Yang(2010) let q tend to 1 as n approaches infinity.For a parametric distribution family G {f (x; θ ) : θ },suppose it is closed under the escort transformation (i.e., θ , θ , s.t. f (x; θ ) f (x; θ )(1/q) ). We have a similar property, θ̃ arg maxθ Eθ0 Lq (f (X; θ )), where θ̃ satisfiesf (x; θ̃ ) f (x; θ0 )(1/q)We now understand that, when maximizing the Lq-likelihood,we are essentially finding the escort distribution of the true density, not the true density itself, so our MLqE is asymptoticallybiased. However, this bias can be compensated by variance reduction if the distortion parameter q is properly selected. Takethe MLqE for the normal distribution for example. With an appropriate q 1, the MLqE will partially ignore the data pointson the tails while focusing more on fitting data points aroundthe center. The MLqE obtained this way is possibly biased (especially for the scale parameter), but will be less volatile to asignificant change of data on the tails, hence, a good example ofbias-variance tradeoff. q can be considered as a tuning parameterthat adjusts the magnitude of the bias-variance tradeoff.2.3 Confidence Intervalsi 1where wi ϕ(xi ; μ̂MLqE , σˆ2 MLqE )1 q and ϕ is a normal probability density function.From Equations (2) and (3), we conclude that the MLqE ofthe mean and the variance of a normal distribution are just theweighted mean and weighted variance. When q 1, the MLqEgives smaller weights for data points lying in the tail of thenormal distribution, and puts more weights on data points nearthe center. By doing so, the MLqE becomes less sensitive tooutliers than the MLE at the cost of introducing bias into theestimation. A simple and fast reweighting algorithm is availableThere are generally two ways to construct confidence intervals for the MLqE. One is parametric, the other is nonparametric.In this section, we discuss the univariate case. The multivariatecase can be extended naturally.For the parametric way, we know that the MLqE is an Mestimator, whose asymptotic variance is available. To have theasymptotic variance be valid, we need the sample size to bereasonably large so that the central limit theorem works. However, in our application, the MLqE deals with small or moderatesample sizes in most cases. So the parametric way is not ideal,but it does provide a guideline to evaluate the estimator.The second way is the nonparametric bootstrap method. Wecreate bootstrap samples from the original sample, calculate

916Journal of the American Statistical Association, September 2013their MLqEs for all bootstrap samples. We further calculatethe lower and upper quantiles of these MLqEs, and call thesequantiles the lower and upper bounds of the confidence interval.This method is model agnostic, and works well with the MLqE.Downloaded by [Johns Hopkins University] at 14:19 23 October 20133. MLq E OF MIXTURE MODELSWe now look at the problem of estimating mixture models. A mixture model is defined as f (x) kj 1 πj fj (x; θj ).Unlike the exponential family, which is proved to be closedunder the escort transformation (Equation (4)), the mixturemodel family is not closed under such a transformation.For example, consider a mixture model with the complexity k 2. The escort transformation with 1/q 2 of thisdistribution is f (x)(1/q) (π1 ϕ1 (x) π2 ϕ2 (x))2 π12 ϕ1 (x)2 π22 ϕ2 (x)2 2π1 π2 ϕ1 (x)ϕ2 (x), which is a mixture model withthree components.More generally, suppose f0 F, where F is a mixture model(1/q) / F, we know thatfamily with complexity k. Since f0(1/q)f0 g̃ : arg max Ef0 Lq (g(X)),g F(1/q)where g̃ can be considered as the projection of f0onto F.Again, the MLqE of mixture models brings more bias to theestimate. This time, the new bias is a model bias as opposedto the estimation bias which we have discussed in the previoussection. When estimating mixture models using the MLqE, wecarry two types of bias: estimation bias and model bias. Thedistortion parameter q now adjusts both of them. This idea isillustrated in Figure 1(a).There is a simple way to partially correct the bias. Sincewe know that the MLqE is unbiased for the escort distribution of the true distribution. After we obtain the MLqE fromdata, fˆMLqE , we can blow it up by a power transformationqqg fˆMLqE / fˆMLqE dx to get a less biased estimate. However,this only partially corrects the bias since the projection fromthe escort distribution onto the mixture model family cannot berecovered by this transformation.Because the MLqE has the desirable property of being robustagainst outliers, we introduce the gross error model to evaluatethe MLqE’s performance. A gross error model is defined asf0 (x) (1 )f0 (x) ferr (x), where f0 is a mixture modelwith complexity k, ferr can be considered as a measurementerror component, and is the contamination ratio. Hence, f0 is also a mixture model with complexity k 1. The gross errordensity f0 can be considered as a small deviation from the targetdensity f0 . To build an estimator for f0 that is robust againstferr , we apply the MLqE. Generally, there are two ways to applythe MLqE in this situation.First, we can directly use a mixture model with complexityk to estimate f0 based on data from f0 . We call this approachthe direct approach. This time the model is more complex thanbefore. The idea is illustrated in Figure 1(b). Suppose, F is a/ F,mixture model family with complexity k, and f0 F, f0 (1/q)f0 / F. We obtain the MLqE of f0 (x), g̃, by (1/q)f0 g̃ : arg max Ef0 Lq (g(X)).g FFigure 1. Illustration of the MLqE of mixture models: (a) showsthe usual case, which is the MLqE of mixture models with correctlyspecified models, (b) shows the MLqE of nonmeasurement error components f0 within the gross error model f0 using the misspecifiedmodel, and (c) shows the MLqE of nonmeasurement error componentsf0 within the gross error model f0 using the correctly specified model.Here we use the estimation bias and the model bias to offset themeasurement error effect on f0 . Please note that this approachis essentially an estimation under the misspecified model.The second approach is that we use a mixture model withcomplexity k 1 to estimate f0 and project the estimate tothe k component mixture model family by removing the largestvariance component (i.e., the measurement error component)and normalizing the weights. We call this approach the indirectapproach. The projected model is our estimate for f0 . In thiscase, we essentially treat the parameters of the measurementerror component as nuisance parameters. This idea is illustratedin Figure 1(c). In Figure 1(c), g̃ is our estimate of f0 . And g̃0 ,the projection of g̃ onto F0 , is our estimate of f0 . This approachis an estimation conducted under the correctly specified model.Although the model is correctly specified, we may have higherestimation variance as we estimate more parameters.In this article, we will study the MLqE using the above twoapproaches.Please note that, when q 1, the MLqE is an inconsistentestimator. Ferrari and Yang (2010) let q 1 as n to forcethe consistency. In our case, we allow the MLqE to be inconsistent because our data are contaminated. We are no longerafter the true underlying distribution f0 that generates the data,but are more interested in estimating the nonmeasurement errorcomponents f0 using the contaminated data. Since the goal isnot to estimate f0 , being consistent will not help the estimatorin terms of robustness.

Qin and Priebe: Maximum Lq -Likelihood Estimation via the Expectation-Maximization Algorithmn 4. EM ALGORITHM WITH Lq -LIKELIHOODWe now propose a variation of the EM algorithm—the expectation maximization algorithm with Lq-likelihood (EM-Lq),which gives the local maximum Lq-likelihood. Before introducing our EM-Lq, let us briefly review the rationale of the EM.Throughout this article, we use X, Z, Z for random variablesand vectors, and x, z, z for realizations.4.1 Why Does the EM Algorithm Work?The EM algorithm is an iterative method for finding a localmaximum likelihood by making use of observed data X andmissing data Z. The rationale behind the EM is thatn log p(xi ; ) Downloaded by [Johns Hopkins University] at 14:19 23 October 2013i 1n i 1 E oldi 1 p(Z X; old )p(Z X; )Lq (p(Z X; ))p(Z X; old )1 q A( , old ) n E oldn i 1where J ( , old ) is the expected complete log-likelihood,and K( , old ) takes its minimum at old and K( , old ) old 0. Standing at the current estimate old , to climb uphill on ni 1 log p(xi ; ) only requires usto climb J, and K will automatically increase. Meanwhile,the in

Maximum Lq-Likelihood Estimation via the Expectation-Maximization Algorithm: A Robust Estimation of Mixture Models Yichen QIN and Carey E. PRIEBE We introduce a maximum Lq-likelihood estimation (MLqE) of mixture models using our proposed expectation-maximization (EM) al- gorithm, namely the EM algorithm with Lq-likelihood (EM-Lq).

Related Documents:

The Qin and Han Dynasties The Qin also built the terracotta warrior tomb for Shi Huangdi. The Qin ruled as an autocracy. They used harsh punishment and censorship. Discontent grew and in 202 B.C. Liu Bang overthrew the Qin dynasty and established the Han. The Han dynasty took its name from the Han River. Ruled for 400 years.

The process by which the state of Qin came to unify China has long been the subject of much academic study.1 One of the salient achievements of the Qin empire, often mentioned by scholars, was the unification of measurement systems, including lengths, volumes, and weights. Qin Shihuangdi 秦始皇帝 (r. 246–221–210 BCE) and Qin Ershi

_ China under their leadership. ( 221 BCE) Qin Dynasty Tuesday, November 19, 2013. After over 200 years of Þghting the _ was able to _ China under their leadership. ( 221 BCE) Qin Dynasty unify Tuesday, November 19, 2013. The ruler of the Qin Dynasty was _. He became the _

16.2The Government of Imperial China In 221 B.C.R., Prince Zheng, the head of the state of Qin. became the first Chinese ruler to claim the title of emperor. He took the name Qin Shihuangdi. which means "First Emperor of Qin." From that time on, China generally had an imperial government headed by an emperor or, sometimes, an empress.

Classical China: The Qin and Han Dynasties The Qin Dynasty 221-206 B.C. The Han Dynasty 206 B.C.- 200 A.D.

Each of China’s early dynasties was led by rulers who were very different. In this section, you will see how the Qin and Han dynasties differed because of their rulers. Focusing on the Qin Shihuangdi used harsh methods to unify and defend China.(page 241) Developments during the Han d

Each of China’s early dynasties was led by rulers who were very different. In this section, you will see how the Qin and Han dynasties differed because of their rulers. Focusing on the Qin Shihuangdi used harsh methods to unify and defend China.(page 241) Developments during the Han d

The risks of introducing artificial intelligence into national militaries are not small. Lethal autonomous weapon systems (LAWS) receive popular attention because such systems are easily imagined .