avatar

Cheng-Chun Hsu

PhD student
The University of Texas at Austin
chengchun (at) utexas.edu


About Me

I am a Ph.D. student in Computer Science at the University of Texas at Austin. My research lies at the intersection of robotics and computer vision.

Publications [Google Scholar]

  1. Cheng-Chun Hsu, Bowen Wen, Jie Xu, Yashraj Narang, Xiaolong Wang, Yuke Zhu, Joydeep Biswas, Stan Birchfield
    International Conference on Robotics and Automation (ICRA), 2025.

    We enable robots to learn everyday tasks from human video demonstrations by usinbg object-centric representation. By predicting future object pose trajectories, SPOT achieves strong generalization capabilities with only eight human video demonstrations.

  2. Cheng-Chun Hsu, Ben Abbatematteo, Zhenyu Jiang, Yuke Zhu, Roberto Martín-Martín, Joydeep Biswas
    Mobile Manipulation Workshop at ICRA, 2024.
    Spotlight Presentation
    We enable mobile manipulators to perform long-horizon tasks by autonomously exploring and building scene-level articulation models of articulated objects. It maps the scene, infers object properties, and plans sequential interactions for accurate real-world manipulation.

  3. Cheng-Chun Hsu, Zhenyu Jiang, Yuke Zhu
    International Conference on Robotics and Automation (ICRA), 2023.

    We develop an interactive perception approach for robots to build indoor scene articulation models by efficiently discovering and characterizing articulated objects through coupled affordance prediction and articulation inference.

  4. Zhenyu Jiang, Cheng-Chun Hsu, Yuke Zhu
    Computer Vision and Pattern Recognition (CVPR), 2022.
    Oral Presentation
    We develop an approach that builds digital twins of articulated objects by learning their articulation models and 3D geometry from visual observations before and after interaction.

  5. Cheng-Chun Hsu, Yi-Hsuan Tsai, Yen-Yu Lin, and Ming-Hsuan Yang
    European Conference on Computer Vision (ECCV), 2020.

    We propose a domain adaptation framework for object detection that uses pixel-wise objectness and centerness to align features, focusing on foreground pixels for better cross-domain adaptation.

  6. Cheng-Chun Hsu*, Kuang-Jui Hsu*, Chung-Chi Tsai, Yen-Yu Lin, and Yung-Yu Chuang
    Neural Information Processing Systems (NeurIPS), 2019.

    We propose a weakly supervised instance segmentation method that leverages Multiple Instance Learning (MIL) to address ambiguous foreground separation from bounding box annotations.

  7. Shintami Chusnul Hidayati, Cheng-Chun Hsu, Yu-Ting Chang, Kai-Lung Hua, Jianlong Fu, and Wen-Huang Cheng
    ACM International Conference on Multimedia (MM), 2018.
    Oral Presentation
    We propose to learn clothing style and body shape compatibility from social big data, offering personalized outfit recommendations by factoring in a user's body shape.

Technical Reports

  1. Cheng-Chun Hsu*, Kuang-Jui Hsu*, Chiachen Ho, Yen-Yu Lin, and Yung-Yu Chuang
    Technical report, 2019.

    We propose a weakly supervised instance segmentation method using image-level labels, leveraging MIL, semantic segmentation, and a novel refinement module.


© 2022 Cheng-Chun Hsu. Powered by Jekyll with minimal-light theme.