Welcome to the upgraded MacSphere! We're putting the finishing touches on it; if you notice anything amiss, email macsphere@mcmaster.ca

Self-Supervised Masked Autoencoding Meets Federated Learning for Electric Vehicle Battery State-of-Health Estimation

dc.contributor.advisorAhmed, Ryan
dc.contributor.authorIsmail, Mohanad
dc.contributor.departmentMechanical Engineeringen_US
dc.date.accessioned2025-07-16T20:13:02Z
dc.date.available2025-07-16T20:13:02Z
dc.date.issued2025
dc.description.abstractEVs live and die by their batteries. To keep drivers safe and confident in their vehicles, we need efficient, accurate, and private ways to track each battery's SoH. But, EV labelled data is scarce, sharing raw data raises privacy flags, and big models strain on-board hardware. This thesis tackles all three problems through a two-step remedy in one shot. 1. Learn data representations without needing labels: Each car trains a small autoencoder to reconstruct its own collected sensor data after randomly hiding parts of the signal. 2. Share knowledge, not data: Instead of uploading the raw collected data, every car sends only its trained model parameters to a remote cloud server. The server aggregates parameters from all cars and sends the improved model back. Four simple questions guide our work: 1. Does this usage of unlabelled data improve the model's performance? 2. How much of the signal should be hidden to get the best representation learning? 3. What is the optimal strategy for incorporating the limited labelled data available into the model? 4. Does this aggregation of separately trained models hurt accuracy compared with a fully centralized approach? Our experiments show a 17% lower average MAE, with up to a 60% improvement in the best cases, when we make use of the available unlabelled data versus training exclusively on labelled data. Hiding 30-40% of signals strikes the balance between challenge and clarity. Finally, aggregation of models on average stays within 0.05Ah of centralized training, virtually no loss, with zero raw-data exposure. This thesis incorporates cloud computing, SSL, and FL to present a light, privacy-friendly pipeline for fleet-wide SoH estimation, evidence that unfrozen fine-tuning outshines frozen variants, the first systematic look at how masking ratio shapes battery time-series representation learning, and practical proof that sharing model weights instead of data keeps accuracy basically untouched and privacy intact.en_US
dc.description.degreeMaster of Applied Science (MASc)en_US
dc.description.degreetypeThesisen_US
dc.identifier.urihttp://hdl.handle.net/11375/31980
dc.language.isoenen_US
dc.subjectElectric Vehicleen_US
dc.subjectState-of-Health Estimationen_US
dc.subjectSelf-Supervised Learningen_US
dc.subjectMasked Autoencodingen_US
dc.subjectFederated Learningen_US
dc.subjectCloud Computingen_US
dc.subjectEdge Computingen_US
dc.subjectFine-Tuning Strategiesen_US
dc.subjectMasking Ratio Optimizationen_US
dc.subjectData Scarcity and Heterogeneityen_US
dc.subjectPrivacy-Preserving Machine Learningen_US
dc.titleSelf-Supervised Masked Autoencoding Meets Federated Learning for Electric Vehicle Battery State-of-Health Estimationen_US
dc.typeThesisen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
ismail_mohanad_2025july_masc.pdf
Size:
15.37 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.68 KB
Format:
Item-specific license agreed upon to submission
Description: