Welcome to the upgraded MacSphere! We're putting the finishing touches on it; if you notice anything amiss, email macsphere@mcmaster.ca

Explainable Learning of Long-term Dependencies through Machine Learning

dc.contributor.advisorDown, Douglas
dc.contributor.authorMartinez-Garcia, Fernando
dc.contributor.departmentComputing and Softwareen_US
dc.date.accessioned2024-01-29T19:52:01Z
dc.date.available2024-01-29T19:52:01Z
dc.date.issued2024
dc.description.abstractMachine learning-based models have yielded remarkable results in a wide range of applications, revolutionizing industries over the last few decades. However, a variety of challenges from the technical point of view, such as the drastic increase in model size and complexity, have become a barrier for their portability and human interpretation. This work focuses on enhancing specific machine learning models used in the time-series forecasting domain. The study begins by demonstrating the effectiveness of a simple and interpretable-by-design machine learning model in handling a real-world time-series industry-related problem. This model incorporates new data while dynamically forgetting previous information, thus promoting continuous learning and adaptability laying the groundwork for practical applications within industries where real-time interpretable adaptation is crucial. Then, the well-established LSTM Neural Network, an advanced but less interpretable model able to learn long and more complex time dependencies, is modified to generate a model, named E-LSTM, with extended temporal connectivity to better capture long-term dependencies. Experimental results demonstrate improved performance with no significant increase in model size across various datasets, showcasing the potential to have balance between performance and model size. Finally, a new LSTM architecture built upon the E-LSTM’s increased temporal connectivity while embedded with interpretability is proposed, called Generalized Interpretable LSTM (GI-LSTM). This architecture is designed to offer a more holistic interpretation of its learned long-term dependencies, providing semi-local interpretability by offering insights into the detected relevance across time-series data. Furthermore, the GI-LSTM outperforms alternative models, generally produces smaller models, and shows that performance does not necessarily come at the cost of interpretability.en_US
dc.description.degreeDoctor of Philosophy (PhD)en_US
dc.description.degreetypeThesisen_US
dc.description.layabstractMachine learning has made big advances and transformed industries, but challenges such as growing model sizes and diminishing interpretability have hindered their usage and reliability. This research aims to enhance machine learning models for time-series forecasting. It starts by showcasing an interpretable-by-design linear model and its effectiveness in solving a real-world industry-related problem by means of incorporating new data while dynamically forgetting old information. Then, to consider nonlinear time-series components, the study delves into improving the Long Short-Term Memory (LSTM) Neural Network by creating an extended version, named E-LSTM, able to better exploit nonlinear long-term dependencies, resulting in a model of similar size and improved performance. Finally, the Generalized Interpretable LSTM (GI-LSTM), a more general LSTM architecture with higher temporal connectivity and embedded interpretability, is introduced. This architecture is shown to offer a more holistic interpretation of learned long-term dependencies while outperforming the previous architectures, all while keeping a compact model size.en_US
dc.identifier.urihttp://hdl.handle.net/11375/29455
dc.language.isoenen_US
dc.subjectExplainable, Interpretable, AI, RNN, LSTM, Machine Learning, Time series, Adaptive Predictive Control, E-LSTM GI-LSTMen_US
dc.titleExplainable Learning of Long-term Dependencies through Machine Learningen_US
dc.typeThesisen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Martinez_Fernando_2023December_PhD.pdf
Size:
5.89 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.68 KB
Format:
Item-specific license agreed upon to submission
Description: