Skip navigation
  • Home
  • Browse
    • Communities
      & Collections
    • Browse Items by:
    • Publication Date
    • Author
    • Title
    • Subject
    • Department
  • Sign on to:
    • My MacSphere
    • Receive email
      updates
    • Edit Profile


McMaster University Home Page
  1. MacSphere
  2. Open Access Dissertations and Theses Community
  3. Open Access Dissertations and Theses
Please use this identifier to cite or link to this item: http://hdl.handle.net/11375/30425
Title: Online EM for Exponential Family Distributions
Authors: Rahimipour, Shiva
Advisor: McNicholas, Paul D.
Department: Computational Engineering and Science
Keywords: Online learning;Gaussian mixture models;Exponential Family Distributions;Mirror descent algorithm;Principle Component Analysis
Publication Date: 2024
Abstract: The expectation-maximization (EM) algorithm is a widely utilized method for fitting probabilistic models with missing data or latent variables. However, applying the EM algorithm to high-dimensional datasets requires storing the entire dataset in computer memory, which is necessary at each iteration. Fortunately, online learning algorithms are available that overcome this issue by processing data sequentially and updating model parameters at each step. In this thesis, we introduce an approach for online training of mixture models of exponential family distributions through modifications to the EM algorithm. The first part of our contribution is the learning of Gaussian mixture models (GMMs) using alternating mirror descent updates. This method employs custom Bregman divergences for parameter updates and allows for the processing of one unit of data at each iteration. Our second contribution is an improved version of the EM algorithm for exponential family distributions. In this case, we replace the M-step updates with mirror descent updates and further enhance them by introducing a new Bregman divergence for the updates. Finally, we apply the improved EM algorithm to online principal component analysis (PCA) for exponential families, where the number of principal components is adaptively learned by pruning factors with low correlations to the observations. We evaluate the performance of our proposed approaches by comparing them with both simulation and real datasets.
URI: http://hdl.handle.net/11375/30425
Appears in Collections:Open Access Dissertations and Theses

Files in This Item:
File Description SizeFormat 
Rahimipour_Shiva_202409_PhD.pdf
Embargoed until: 2025-09-25
68.05 MBAdobe PDFView/Open
Show full item record Statistics


Items in MacSphere are protected by copyright, with all rights reserved, unless otherwise indicated.

Sherman Centre for Digital Scholarship     McMaster University Libraries
©2022 McMaster University, 1280 Main Street West, Hamilton, Ontario L8S 4L8 | 905-525-9140 | Contact Us | Terms of Use & Privacy Policy | Feedback

Report Accessibility Issue