Root selection in normal mixture models
From MaRDI portal
Publication:693259
DOI10.1016/j.csda.2012.01.022zbMath1252.62013MaRDI QIDQ693259
Publication date: 7 December 2012
Published in: Computational Statistics and Data Analysis (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/j.csda.2012.01.022
62F10: Point estimation
Related Items
Estimation under copula-based Markov normal mixture models for serially correlated data, Finite Mixture of Linear Regression Models: An Adaptive Constrained Approach to Maximum Likelihood Estimation, A simple root selection method for univariate finite normal mixture models, Discussion on “Distributional independent component analysis for diverse neuroimaging modalities” by Ben Wu, Subhadip Pal, Jian Kang, and Ying Guo, Merging components in linear Gaussian cluster-weighted models, Root selection in normal mixture models, Dealing with multiple local modalities in latent class profile analysis, Editorial: The 2nd special issue on advances in mixture models, A multivariate linear regression analysis using finite mixtures of \(t\) distributions, Semiparametric mixture: continuous scale mixture approach, Scale-constrained approaches for maximum likelihood estimation and model selection of clusterwise linear regression models, Density deconvolution under a \(k\)-monotonicity constraint, A note on maximum likelihood estimation for mixture models, Model-based clustering with determinant-and-shape constraint, Maximizing leave-one-out likelihood for the location parameter of unbounded densities, Mixtures of multivariate contaminated normal regression models, Robust clustering in regression analysis via the contaminated Gaussian cluster-weighted model, Eigenvalues and constraints in mixture modeling: geometric and computational issues, A data driven equivariant approach to constrained Gaussian mixture modeling, Probabilistic assessment of model-based clustering, Assessment of the number of components in Gaussian mixture models in the presence of multiple local maximizers, The Mixturegram: A Visualization Tool for Assessing the Number of Components in Finite Mixture Models
Uses Software
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Methods for merging Gaussian mixture components
- Root selection in normal mixture models
- Bayesian regularization for normal mixture estimation and model-based clustering
- Constrained monotone EM algorithms for finite mixture of multivariate Gaussians
- Inference for multivariate normal mixtures
- A constrained formulation of maximum-likelihood estimation for normal mixture distributions
- Nonparametric maximum likelihood estimation by the method of sieves
- The distribution of the likelihood ratio for mixtures of densities from the one-parameter exponential family
- Degeneracy in the maximum likelihood estimation of univariate Gaussian mixtures with EM.
- A computational strategy for doubly smoothed MLE exemplified in the normal mixture model
- Markov chain Monte Carlo methods and the label switching problem in Bayesian mixture modeling
- Types of likelihood maxima in mixture models and their implication on the performance of tests
- Strong consistency of the maximum likelihood estimator for finite mixtures of location-scale distributions when the scale parameters are exponentially small
- Consistency of the Maximum Likelihood Estimator in the Presence of Infinitely Many Incidental Parameters
- Penalized Maximum Likelihood Estimator for Normal Mixtures
- Bayesian Mixture Labeling by Highest Posterior Density
- Note on the Consistency of the Maximum Likelihood Estimate