Please use this identifier to cite or link to this item:
http://hdl.handle.net/11375/32482
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.advisor | Kratsios, Anastasis | - |
dc.contributor.author | Arabpour Dahoei, Reza | - |
dc.date.accessioned | 2025-10-02T16:16:51Z | - |
dc.date.available | 2025-10-02T16:16:51Z | - |
dc.date.issued | 2025-11 | - |
dc.identifier.uri | http://hdl.handle.net/11375/32482 | - |
dc.description.abstract | This thesis presents two significant research contributions: one focuses on improving the adaptation of large language models (LLMs) using parameter-efficient fine-tuning (PEFT), and the other addresses the effective modelling of history-dependent stochastic processes—specifically Volterra processes, which are commonly applied in quantitative finance. In the first part, I introduce a user-friendly adaptation pipeline that boosts the performance of a standard foundation model, bringing it much closer to a fully fine-tuned, task-specific version. Remarkably, it achieves this while using significantly less compute and memory, all while keeping data private. The pipeline leverages existing learnable low-rank adapters (LoRA) for known datasets and predicts adapter values for new datasets using this readily available information. Its main advantage is that it can run on a standard laptop without requiring GPU power, ensuring that data remains local. This method effectively closes about half of the performance gap between an untuned base model and a fully fine-tuned one, making specialized models more accessible to researchers, practitioners, and everyday users who lack expensive infrastructure or work with sensitive data on devices like smartphones. The second part addresses a computational challenge in translating the non-Markovian Volterra process into a format suitable for computation. This translation is difficult because the data history dimension affecting the current state grows with the length of the path. I propose a two-step approach to make this process manageable: first, the Volterra process is mapped onto a simpler, lower-dimensional manifold; then, a geometric deep learning model—a "hypernetwork"—is applied, specifically designed for the manifold’s structure. We provide both mathematical and computational evidence demonstrating the model’s effectiveness and practicality (with proofs developed by co-authors available in the main paper), along with extensive testing of each parameter to validate our approach. | en_US |
dc.language.iso | en | en_US |
dc.subject | Machine Learning | en_US |
dc.subject | Deep Learning | en_US |
dc.subject | Geometric Deep Learning | en_US |
dc.subject | Financial Mathematics | en_US |
dc.subject | Time Series | en_US |
dc.subject | Financial Time Series | en_US |
dc.subject | Foundation Models | en_US |
dc.subject | Large Language Models | en_US |
dc.subject | LLMs | en_US |
dc.subject | Hypernetworks | en_US |
dc.subject | Mathematics | en_US |
dc.subject | Artificial Intelligence | en_US |
dc.title | Geometric Deep Learning For Financial Time Series and Efficient Fine-Tuning of Foundation Models | en_US |
dc.title.alternative | Geometric Deep Learning for Time Series and Foundation Models | en_US |
dc.type | Thesis | en_US |
dc.contributor.department | Computational Engineering and Science | en_US |
dc.description.degreetype | Thesis | en_US |
dc.description.degree | Master of Science (MSc) | en_US |
dc.description.layabstract | This thesis presents two contributions at the intersection of artificial intelligence and mathematics. First, I introduce a novel method for adapting large language models on widely available hardware. This approach recovers half of the performance lost when using an untuned base model instead of a GPU fine-tuned one, while running on a single laptop with minimal cost and energy consumption. It makes specialized models more accessible, preserves privacy by keeping data local, and promotes environmentally responsible computing. Second, I develop a practical framework for working with history-dependent stochastic processes commonly used in quantitative finance. Such processes are often too large to compute efficiently. The method proposed here compresses them into a low-dimensional representation and then applies a computational model, enabling efficient simulation, estimation, and practical application. Together, these contributions introduce novel algorithms capable of addressing real-world problems from fresh perspectives. | en_US |
Appears in Collections: | Open Access Dissertations and Theses |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
ArabpourDahoei_Reza_202509_MSc.pdf | 3.67 MB | Adobe PDF | View/Open |
Items in MacSphere are protected by copyright, with all rights reserved, unless otherwise indicated.