Convergence analysis for a multiplicatively relaxed EM algorithm
- 1 October 1991
- journal article
- research article
- Published by Wiley in Mathematical Methods in the Applied Sciences
- Vol. 14 (8) , 573-593
- https://doi.org/10.1002/mma.1670140805
Abstract
The expectation maximization (EM) algorithm is an iterative procedure used to determine maximum likelihood estimators in situations of incomplete data. In the case of independent Poisson variables it converges to a solution of a problem of the form min ∑[〈ai,x〉 −bilog 〈ai,x〉] such thatx⩾0. Thus, it can be used to solve systems of the formAx=b,x⩾0 (withAstochastic andbpositive.) It converges to a feasible solution if it exists and to an approximate one otherwise (the one that minimizesd(b,Ax), wheredis the Kullback–Leibler information divergence). We study the convergence of the multiplicatively relaxed version proposed by Tanaka for use in positron emission tomography. We prove global convergence in the underrelaxed and unrelaxed cases. In the overrelaxed case we present a local convergence theorem together with two partial global results: the sequence generated by the algorithm is bounded and, if it converges, its limit point is a solution of the aforementioned problem.Keywords
This publication has 6 references indexed in Scilit:
- A Fast Reconstruction Algorthm for Stationary Positron Emission Tomography Based on a Modified EM AlgorithmIEEE Transactions on Medical Imaging, 1987
- A Statistical Model for Positron Emission TomographyJournal of the American Statistical Association, 1985
- Maximum Likelihood from Incomplete Data Via the EM AlgorithmJournal of the Royal Statistical Society Series B: Statistical Methodology, 1977
- Second Order Conditions for Constrained MinimaSIAM Journal on Applied Mathematics, 1967
- Proximity Maps for Convex SetsProceedings of the American Mathematical Society, 1959
- Proximity maps for convex setsProceedings of the American Mathematical Society, 1959