Please use this identifier to cite or link to this item:
http://hdl.handle.net/11375/29455
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.advisor | Down, Douglas | - |
dc.contributor.author | Martinez-Garcia, Fernando | - |
dc.date.accessioned | 2024-01-29T19:52:01Z | - |
dc.date.available | 2024-01-29T19:52:01Z | - |
dc.date.issued | 2024 | - |
dc.identifier.uri | http://hdl.handle.net/11375/29455 | - |
dc.description.abstract | Machine learning-based models have yielded remarkable results in a wide range of applications, revolutionizing industries over the last few decades. However, a variety of challenges from the technical point of view, such as the drastic increase in model size and complexity, have become a barrier for their portability and human interpretation. This work focuses on enhancing specific machine learning models used in the time-series forecasting domain. The study begins by demonstrating the effectiveness of a simple and interpretable-by-design machine learning model in handling a real-world time-series industry-related problem. This model incorporates new data while dynamically forgetting previous information, thus promoting continuous learning and adaptability laying the groundwork for practical applications within industries where real-time interpretable adaptation is crucial. Then, the well-established LSTM Neural Network, an advanced but less interpretable model able to learn long and more complex time dependencies, is modified to generate a model, named E-LSTM, with extended temporal connectivity to better capture long-term dependencies. Experimental results demonstrate improved performance with no significant increase in model size across various datasets, showcasing the potential to have balance between performance and model size. Finally, a new LSTM architecture built upon the E-LSTM’s increased temporal connectivity while embedded with interpretability is proposed, called Generalized Interpretable LSTM (GI-LSTM). This architecture is designed to offer a more holistic interpretation of its learned long-term dependencies, providing semi-local interpretability by offering insights into the detected relevance across time-series data. Furthermore, the GI-LSTM outperforms alternative models, generally produces smaller models, and shows that performance does not necessarily come at the cost of interpretability. | en_US |
dc.language.iso | en | en_US |
dc.subject | Explainable, Interpretable, AI, RNN, LSTM, Machine Learning, Time series, Adaptive Predictive Control, E-LSTM GI-LSTM | en_US |
dc.title | Explainable Learning of Long-term Dependencies through Machine Learning | en_US |
dc.type | Thesis | en_US |
dc.contributor.department | Computing and Software | en_US |
dc.description.degreetype | Thesis | en_US |
dc.description.degree | Doctor of Philosophy (PhD) | en_US |
dc.description.layabstract | Machine learning has made big advances and transformed industries, but challenges such as growing model sizes and diminishing interpretability have hindered their usage and reliability. This research aims to enhance machine learning models for time-series forecasting. It starts by showcasing an interpretable-by-design linear model and its effectiveness in solving a real-world industry-related problem by means of incorporating new data while dynamically forgetting old information. Then, to consider nonlinear time-series components, the study delves into improving the Long Short-Term Memory (LSTM) Neural Network by creating an extended version, named E-LSTM, able to better exploit nonlinear long-term dependencies, resulting in a model of similar size and improved performance. Finally, the Generalized Interpretable LSTM (GI-LSTM), a more general LSTM architecture with higher temporal connectivity and embedded interpretability, is introduced. This architecture is shown to offer a more holistic interpretation of learned long-term dependencies while outperforming the previous architectures, all while keeping a compact model size. | en_US |
Appears in Collections: | Open Access Dissertations and Theses |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Martinez_Fernando_2023December_PhD.pdf | 6.03 MB | Adobe PDF | View/Open |
Items in MacSphere are protected by copyright, with all rights reserved, unless otherwise indicated.