Research
I am broadly interested in robotics, machine learning and reinforcement learning. My research goal is to develope robots capable of performing complex tasks at or beyond human-level proficiency. I am particularly interested in the following areas:
-
Adaptation to distribution shift (e.g., dc-unida, ssd): For robots to be readily used in the real world, they must have almost perfect success rate (> 99%). This requires the ability to adapt (or safely explore) in environments with distribution shifts. I am interested in algorithms that enable test-time adaptation/exploration.
-
Dexterous manipulation: To fully unlock the potential of robots in complex tasks such as in-hand manipulation and tool use, fine-grained control is essential. I am interested in how to scale up RL/IL for these tasks.
|
|
Soft Separation and Distillation: Toward Global Uniformity in Federated Unsupervised Learning
International Conference on Computer Vision (ICCV), 2025
We explore how to improve representation quality (uniformity) under limited and restricted information due to the nature of federated learning (non-IID, features not shared across clients). We improve inter-client uniformity by softly separating client embeddings without disrupting the underlying data structure.
|
|
Tackling Dimensional Collapse toward Comprehensive Universal Domain Adaptation
International Conference on Machine Learning (ICML), 2025
We found that partial alignment loss fails to outperform the simplest baseline—training only on source data—when there is a source-target label imbalance, due to dimensional collapse in target representations. We address this using de-collapse techniques from self-supervised learning, advancing toward more comprehensive universal domain adaptation.
|
|
Open-domain Conversational Question Answering with Historical Answers
Asian Chapter of the Association for Computational Linguistics (AACL), 2022
We propose combining the signal from historical answers with the noise-reduction ability of knowledge distillation to improve information retrieval and question answering.
|
|
Zero-shot Text Behavior Retrieval
Course Project of "Reinforcement Learning", Fall 2023
We propose a method for retrieving task-relevant data for imitation learning without requiring expert demonstrations. Our approach leverages text descriptions in combination with a vision-language model to enable zero-shot behavior retrieval.
|
|
Integrating Self-supervised Speech Model with Pseudo Word-level Targets from Visually-grounded Speech Model
ICASSP Workshop on Self-supervision in Audio, Speech and Beyond, 2024
We propose using vision as a surrogate for paired transcripts to enrich the semantic information in self-supervised speech models.
|
This template is adapted from here.
|
|