Research
I am broadly interested in robotics, machine learning and reinforcement learning. My research goal is to develope robots capable of performing complex tasks at or beyond human-level proficiency. I am particularly interested in the following areas:
-
Adaptation to distribution shift (e.g., DC-UniDA, SSD): For robots to be readily used in the real world, they must have almost perfect success rate (> 99%). This requires the ability to adapt (or safely explore) in environments with distribution shifts. I am interested in algorithms that enable test-time adaptation/exploration.
-
Dexterous manipulation: To fully unlock the potential of robots in complex tasks such as in-hand manipulation and tool use, fine-grained control is essential. I am interested in how to scale up RL/IL for these tasks.
|
|
Soft Separation and Distillation: Toward Global Uniformity in Federated Unsupervised Learning
International Conference on Computer Vision (ICCV), 2025
We explore how to improve representation uniformity under highly non-IID client distributions where representations are non-shared (hence no direct cross-client separation loss). We propose soft separation, which regularizes each client toward a distinct subspace to improve inter-client uniformity while preserving structure.
|
|
Tackling Dimensional Collapse toward Comprehensive Universal Domain Adaptation
International Conference on Machine Learning (ICML), 2025
We found that partial alignment loss fails to outperform the simplest baseline—training only on source data—when there is a source-target label imbalance, due to dimensional collapse in target representations. We address this using de-collapse techniques from self-supervised learning, advancing toward more comprehensive universal domain adaptation.
|
|
Open-domain Conversational Question Answering with Historical Answers
Asian Chapter of the Association for Computational Linguistics (AACL), 2022
We propose combining the signal from historical answers with the noise-reduction ability of knowledge distillation to improve information retrieval and question answering.
|
|
Integrating Self-supervised Speech Model with Pseudo Word-level Targets from Visually-grounded Speech Model
ICASSP Workshop on Self-supervision in Audio, Speech and Beyond, 2024
We propose using vision as a surrogate for paired transcripts to enrich the semantic information in self-supervised speech models.
|
|
Zero-shot Text Behavior Retrieval
Course Project of "Reinforcement Learning", Fall 2023
We propose a method for retrieving task-relevant data for imitation learning without requiring expert demonstrations. Our approach leverages text descriptions in combination with a vision-language model to enable zero-shot behavior retrieval.
|
This template is adapted from here.
|
|