| 
              
              | - [June 2025] 
                Our paper on improving generalization under non-IID data has been accepted to ICCV 2025. See you in Hawaii! |    
              | - [June 2025] 
                I'm fortunate to visit ILIAD at Stanford, hosted by Prof. Dorsa Sadigh. |    
              | - [May 2025] 
                Our paper on universal domain adaptation across any class priors has been accepted to ICML 2025. See you in Vancouver! |  
              
              | Research
                  
                  I am broadly interested in robotics, representation learning and reinforcement learning. My goal is to understand and develop robots with generalist intelligence that can autonomously adapt in the real world and ultimately perform tasks in our daily lives. I am particularly interested in:
                   
                    
                    Representation learning for generalist policies: (1) learning generalized representations from heterogeneous data (2) discovering the underlying shared structure across different modalities (e.g., vision, text, control, touch).
                    
                    
                      Representation learning for adaptation: (1) enabling robust adaptation to distribution shift (2) developing unsupervised reinforcement learning algorithms for rapid online exploration.
                    
                      Dexterous manipulation: scaling up dexterous hand capabilities across diverse tasks by learning from diverse sources (e.g., simulation, human videos).
                     |  
          
            |  | DexDrummer: In-Hand, Contact-Rich, and Long-Horizon Dexterous Robot Drumming 
              Under Review
               
                
               |  
          
            |  | Learning Skill Abstraction from Action-Free Videos 
              ICML Workshop on Building Physically Plausible World Models, 2025
               
                
               |  
            
              |  | Soft Separation and Distillation: Toward Global Uniformity in Federated Unsupervised Learning 
                International Conference on Computer Vision (ICCV), 2025
                 
                  We explore how to improve generalization under highly non-IID client distributions where representations are non-shared. We propose soft separation, which regularizes each client toward a distinct subspace to improve inter-client uniformity (maximizing information) while preserving semantic alignment.
                 |  
              |  | Tackling Dimensional Collapse toward Comprehensive Universal Domain Adaptation 
                International Conference on Machine Learning (ICML), 2025
                 
                  We study how to adapt to arbitrary target domains without assuming any class-set priors. Existing methods suffer from severe negative transfer under large class-set shifts due to the overestimation of importance weights. We propose a simple uniformity loss that increases the entropy of target representations and improves performance across all class-set priors.
                 |  
              |  | Open-domain Conversational Question Answering with Historical Answers 
                Asian Chapter of the Association for Computational Linguistics (AACL), 2022
                 
                  We propose combining the signal from historical answers with the noise-reduction ability of knowledge distillation to improve information retrieval and question answering.
                 |  
          
            |  | Integrating Self-supervised Speech Model with Pseudo Word-level Targets from Visually-grounded Speech Model  
              ICASSP Workshop on Self-supervision in Audio, Speech and Beyond, 2024
                
              Course Project of "Deep Learning for Human Language Processing", Spring 2023
               
                We propose using vision as a surrogate for paired transcripts to enrich the semantic information in self-supervised speech models.
               |  
            |  | Zero-shot Text Behavior Retrieval 
              Course Project of "Reinforcement Learning", Fall 2023
               
                We propose a method for retrieving task-relevant data for imitation learning without requiring expert demonstrations. Our approach leverages text descriptions in combination with a vision-language model to enable zero-shot behavior retrieval.
               |  
            
              | 
 
                  This template is adapted from here.
                 |  |