Please use this identifier to cite or link to this item:
http://hdl.handle.net/11375/32619| Title: | Text-driven Motion Synthesis and Interaction Generation using Masked Deconstructed Diffusion and Multi-task Scene-aware Models |
| Authors: | Chen, Jia |
| Advisor: | Wang, Yingying |
| Department: | Computing and Software |
| Publication Date: | 2025 |
| Abstract: | This thesis introduces a new generative AI approach that addresses three long-standing hurdles in human motion generation: accuracy, speed, and reliable alignment with user-written text. From a simple sentence, the system quickly produces natural, high-quality 3D movements that can be retargeted to digital characters for animation, virtual reality (VR), and games. The experiment demonstrates its practical value in VR, where the generated motions enhance immersion and responsiveness. Building on this, the thesis explores a second, scene-aware model that works with large language models to understand both the instruction and the surrounding scene. It can break down long requests into smaller steps and generate motions that interact with objects, for example walking to a chair and then sitting down. Together, these contributions point to more intuitive, text-driven tools for creating lifelike character animation. |
| URI: | http://hdl.handle.net/11375/32619 |
| Appears in Collections: | Open Access Dissertations and Theses |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| Chen_Jia_2025October_MSc.pdf | 25.97 MB | Adobe PDF | View/Open |
Items in MacSphere are protected by copyright, with all rights reserved, unless otherwise indicated.
