Haau-Sing Xiaocheng Li
PhD
Technical University of Darmstadt (TU Darmstadt)
Interactive Multimodal Learning

Multimodal learning contains a set of challenging tasks. It requires a deep understanding of all relevant modalities (language, vision, software program) and of the relationships between them. Furthermore, the lack of high-quality datasets increases the difficulty of related tasks. In this project, we will work with multimodal data in an interactive setting, which is consistent with the nature of many real-world applications. We will apply this setting to multimodal tasks including question answering, dialogue, and program prediction. We hope that within our setting, models will be able to integrate information of modalities with high human acceptability or executability by recovering weak information from the data. Ideally, we hope our research will turn into real-world applications for human-assistance or educational purposes.

Track:
Academic Track
PhD Duration:
July 1st, 2021 - May 30th, 2025
First Exchange:
January 1st, 2024 - July 1st, 2024
ELLIS Edge Newsletter
Join the 6,000+ people who get the monthly newsletter filled with the latest news, jobs, events and insights from the ELLIS Network.