News
- 07/2024 NEW Invited talk at IROS24 AI Meets Autonomy: Vision, Language, and Autonomous Systems workshop!
- 07/2024 NEW Four papers are accpeted by ECCV 2024, including SceneVerse, SlotLifter, F-HOI, and PQ3D!
- 07/2024 NEW I will direct the Joint Lab of Embodied AI and Humanoid Robot between and BIGAI and UniTree!
- 06/2024 NEW Host the 1st workshop on New Trends in Multimodal Human Action Perception, Understanding and Generation at CVPR24!
- 03/2024 NEW Four papers are accpeted by CVPR 2024, including three highlight papers! Three of them are about human motion modeling and skill learning, one is about 3D scene synthesis for embodied AI!
- Invited talk at DeepMind Robotics to introduce LEO
- 06/2023 Three papers are accepeted by ICCV 2023!
- 06/2023 We are organizing the 4th CVPR 2023 workshop: 3D Scene Understanding for Vision, Graphics and Robotics, welcome to join us!
- 01/2023 Two papers including SceneDiffuser and GAPartNet are accepted by CVPR 2023! SceneDiffuser code are released!
- 01/2023 Three papers are accepted by ICLR 2023 with one spotlight!
- 01/2023 GenDexGrasp is accepted by ICRA 2023. It proposes a generalizable dexterous grasping model for robotics!
- 10/2021 Invited talk about Compositional Strucutres in Vision and Language at StruCo3D2021 workshop.
- 06/2021 I am co-organizing the 3rd CVPR 2021 workshop: 3D Scene Understanding for Vision, Graphics and Robotics.
- 06/2021 Defend my Ph.D. dissertation Human-like Holistic 3D Scene Understanding.
- 07/2020 Our paper won the Best Paper Award in ICML2020 Workshop on Bridge Between Perception and Reasoning: Graph Neural Networks & Beyond.
- 07/2020 Two papers accepted by ECCV 2020, one of them as oral presentation.
- 05/2020 Awarded the UCLA Dissertation Year Fellowship.
- 03/2020 I am co-organizing the 2nd CVPR 2020 workshop: 3D Scene Understanding for Vision, Graphics and Robotics.
Teaching
- Computer Vision (PKU, Fall 2022), from a modern view, co-teach with Yixin Zhu
- Early and Mid-level Computer Vision (PKU, Fall 2022, Fall 2023), from a statistical and Marr's view, co-teach with Song-Chun Zhu
Preprints
|
Closed-Loop Open-Vocabulary Mobile Manipulation with GPT-4V
Peiyuan Zhi*,
Zhiyuan Zhang*,
Muzhi Han,
Zeyu Zhang, Zhitian Li,
Ziyuan Jiao,
Baoxiong Jia,
Siyuan Huang
Arxiv
Paper / Project
We present COME-robot, the first closed-loop framework utilizing the GPT-4V vision-language foundation model for open-ended reasoning and adaptive planning in real-world scenarios.
|
|
Task-oriented Sequential Grounding in 3D Scenes
Zhuofan Zhang,
Ziyu Zhu,
Pengxiang Li,
Tengyu Liu,
Xiaojian Ma,
Yixin Chen,
Baoxiong Jia,
Siyuan Huang,
Qing Li
Arxiv
Paper / Project / Code
We propose a new task: Task-oriented Sequential Grounding in 3D scenes, wherein an agent must follow detailed step-by-step instructions to complete daily activities by locating a sequence of target objects in indoor scenes.
|
Thesis
|
Human-like Holistic 3D Scene Understanding
Siyuan Huang
UCLA, 2021
Ph.D. Thesis
|
Publications
|
PhyRecon: Physically Plausible Neural Scene Reconstruction
Junfeng Ni*,
Yixin Chen*, Bohan Jing,
Nan Jiang,
Bin Wang,
Bo Dai,
Puhao Li,
Yixin Zhu,
Song-Chun Zhu,
Siyuan Huang
NeurIPS 2024
Paper / Project / Code
we introduce PhyRecon, the first approach to leverage both differentiable rendering and differentiable physics simulation to learn implicit surface representations for neural scene reconstruction.
|
|
Multi-modal Situated Reasoning in 3D Scenes
Xiongkun Linghu*,
Jiangyong Huang*,
Xuesong Niu,
Xiaojian Ma,
Baoxiong Jia†,
Siyuan Huang†
NeurIPS 2024 (Dataset and Benchmark Track)
Paper / Project / Code
we propose Multi-modal Situated Question Answering (MSQA), a large-scale multi-modal situated reasoning dataset, scalably collected leveraging 3D scene graphs and vision-language models (VLMs) across a diverse range of real-world 3D scenes.
|
|
SlotLifter: Slot-guided Feature Lifting for Learning Object-centric Radiance Fields
Yu Liu*,
Baoxiong Jia*,
Yixin Chen,
Siyuan Huang
ECCV 2024
Paper / Project
We propose SlotLifter, a novel object-centric radiance model addressing scene reconstruction and decomposition jointly via slot-guided feature lifting..
|
|
Unifying 3D Vision-Language Understanding Via Promptable Queries
Ziyu Zhu,
Zhuofan Zhang,
Xiaojian Ma,
Xuesong Niu,
Yixin Chen,
Baoxiong Jia,
Zhidong Deng,
Siyuan Huang†,
Qing Li†
ECCV 2024
Paper / Project / Code
We introduce PQ3D, a unified model capable of using Promptable Queries to tackle a wide range of 3D-VL tasks, from low-level instance segmentation to high-level reasoning and planning.
|
|
F-HOI: Toward Fine-grained Semantic-Aligned 3D Human-Object Interactions
Jie Yang*,
Xuesong Niu*,
Nan Jiang*,
Ruimao Zhang†,
Siyuan Huang†,
ECCV 2024
Paper / Project
We propose a unified model called F-HOI, designed to leverage multimodal instructions and empower the Multi-modal Large Language Model to efficiently handle diverse HOI tasks..
|
|
An Embodied Generalist Agent in 3D World
Jiangyong Huang*,
Silong Yong*,
Xiaojian Ma*,
Xiongkun Linghu*,
Puhao Li,
Yan Wang,
Qing Li,
Song-Chun Zhu,
Baoxiong Jia,
Siyuan Huang
ICML 2024
Paper / Project / Code
We introduce LEO, an embodied multi-modal and multi-task generalist agent that excels in perceiving, grounding, reasoning, planning, and acting in the 3D world.
|
|
PHYSCENE: Physically Interactable 3D Scene Synthesis for Embodied AI
Yangdan Yang*,
Baoxiong Jia*,
Peiyuan Zhi,
Siyuan Huang
CVPR 2024
Highlight, 2.8%
Paper / Project
We introduce PhyScene, a novel method dedicated to generating interactive 3D scenes characterized by realistic layouts, articulated objects, and rich physical interactivity tailored for embodied agents..
|
|
Move as You Say, Interact as You Can:
Language-guided Human Motion Generation with Scene Affordance
Zan Wang,
Yixin Chen,
Baoxiong Jia,
Puhao Li,
Jinlu Zhang, Jinze Zhang,
Tengyu Liu,
Yixin Zhu† ,
Wei Liang†,
Siyuan Huang†
CVPR 2024
Highlight, 2.8%
Paper / Project / Code
We introduce a novel two-stage framework that employs scene affordance as an intermediate representation, effectively linking 3D scene grounding and conditional motion generation.
|
|
Scaling Up Dynamic Human-Scene Interaction Modeling
Nan Jiang*, Zhiyuan Zhang*, Hongjie Li, Xiaoxuan Ma,
Zan Wang,
Yixin Chen,
Tengyu Liu,
Yixin Zhu† ,
Siyuan Huang †
CVPR 2024
Highlight, 2.8%
Paper / Project / Live Demo / Code
We introduce the TRUMANS dataset as the most comprehensive motion-captured HSI dataset currently available, encompassing over 15 hours of human interactions across 100 indoor scenes. We also present the first model that scales up human-scene interaction modeling and achieves remarkable performance in generation.
|
|
AnySkill: Learning Open-Vocabulary Physical Skill for Interactive Agents
Jiemin Cui*,
Tengyu Liu*,
Nian Liu*,
Yaodong Yang,
Yixin Zhu† ,
Siyuan Huang†
CVPR 2024
Paper / Project / Code
We propose AnySkill, a novel hierarchical method that learns physically plausible interactions following open-vocabulary instructions. AnySkill is the first method capable of open-vocabulary physical skill learning for interactive humanoid agents.
|
|
Neural-Symbolic Recursive Machine for Systematic Generalization
Qing Li,
Yixin Zhu,
Yitao Liang,
Ying Nian Wu,
Song-Chun Zhu,
Siyuan Huang
ICLR 2024
Paper
We propose Neural-Symbolic Recursive Machine (NSR) for learning compositional rules from limited data and applying them to unseen combinations in various domains. NSR achieves 100% generalization accuracy on SCAN and
PCFG and outperforms state-of-the-art models on HINT by about 23%.
|
|
Single-view 3D Scene Reconstruction with High-fidelity Shape and Texture
Yixin Chen*,
Junfeng Ni*,
Nan Jiang, Yaowei Zhang,
Yixin Zhu,
Siyuan Huang
3DV 2024
Paper /
Project /
Code
We introduce MultiGrasp, a two-stage method for multi-object grasping on a tabletop with a multi-finger dexterous hand. It involves (i) generating pre-grasp proposals and (ii) executing the grasp and lifting the objects.
|
|
ARNOLD: A Benchmark for Language-Grounded Task Learning with Continuous States in Realistic Scenes
Ran Gong*,
Jiangyong Huang*,
Yizhou Zhao,
Haoran Geng,
Xiaofeng Gao,
Qingyang Wu,
Wensi Ai,
Ziheng Zhou,
Demetri Terzopoulos,
Song-Chun Zhu,
Baoxiong Jia,
Siyuan Huang
ICCV 2023
CoRL 2022 Workshop on Language and Robot Learning
Spotlight Presentation
Arxiv / Project / Code
We present ARNOLD, a
benchmark that evaluates language-grounded task learning with continuous states
in realistic 3D scenes. ARNOLD consists of 8 language-conditioned tasks that
involve understanding object states and learning policies for continuous goals.
|
|
Full-Body Articulated Human-Object Interaction
Nan Jiang*,
Tengyu Liu*,
Zhexuan Cao,
Jieming Cui,
Zhiyuan Zhang,
He Wang,
Yixin Zhu† ,
Siyuan Huang†
ICCV 2023
Paper / Project / Code
We present CHAIRS, a large-scale motion-captured f-AHOI dataset, consisting of 16.2 hours of versatile interactions between 46 participants and 74 articulated and rigid sittable objects. CHAIRS provides 3D meshes of both humans and articulated objects during the entire interactive process, as well as realistic and physically plausible full-body interactions.
|
|
Diffusion-based Generation, Optimization, and Planning in 3D Scenes
Siyuan Huang*,
Zan Wang*,
Puhao Li,
Baoxiong Jia,
Tengyu Liu,
Yixin Zhu,
Wei Liang,
Song-Chun Zhu
CVPR 2023
Paper / Project / Code
We introduce SceneDiffuser, a conditional generative model for 3D scene understanding. SceneDiffuser provides a unified model for solving scene-conditioned generation, optimization, and planning. In contrast to prior work, SceneDiffuser is intrinsically scene-aware, physics-based, and goal-oriented.
|
|
GAPartNet: Learning Generalizable and Actionable Parts for Cross-Category Object Perception and Manipulation
Haoran Geng*, Helin Xu*, Chengyang Zhao*,
Chao Xu,
Li Yi,
Siyuan Huang,
He Wang
CVPR 2023
Highlight, 2.5%
Paper / Project / Code
We propose to learn generalizable object perception and manipulation skills via Generalizable and Actionable Parts,
and present GAPartNet, a large-scale interactive dataset with rich part annotations.
|
|
A Minimalist Dataset for Systematic Generalization of Perception, Syntax, and Semantics
Qing Li,
Siyuan Huang,
Yining Hong,
Yixin Zhu,
Ying Nian Wu,
Song-Chun Zhu
ICLR 2023
Spotlight, 5.7%
Paper
we present a new dataset, HINT, to study machines' capability of learning generalizable concepts at three different levels: perception, syntax, and semantics.
|
|
SQA3D: Situated Question Answering in 3D Scenes
Xiaojian Ma,
Silong Yong,
Zilong Zheng,
Qing Li ,
Yitao Liang,
Song-Chun Zhu,
Siyuan Huang
ICLR 2023
Paper / Project / Code / Slides
We propose a new task to benchmark scene understanding of embodied agents: Situated Question Answering in 3D Scenes (SQA3D). Given a scene context (e.g., 3D scan), SQA3D requires the tested agent to first understand its situation (position, orientation, etc.) in the 3D scene as described by text, then reason about its surrounding environment and answer a question under that situation.
|
|
Improving Object-centric Learning with Query Optimization
Baoxiong Jia*,
Yu Liu*,
Siyuan Huang
ICLR 2023
Paper / Project / Code
Our model, Bi-level Optimized Query Slot Attention, achieves state-of-the-art results on 3 challenging synthetic and 7 complex real-world datasets in unsupervised
image segmentation and reconstruction, outperforming previous baselines by a
large margin.
|
|
HUMANISE: Language-conditioned Human Motion Generation in 3D Scenes
Zan Wang,
Yixin Chen,
Tengyu Liu,
Yixin Zhu,
Wei Liang,
Siyuan Huang
NeurIPS 2022
Paper / Project / Code
We present a novel scene-and-language conditioned generative model that can produce 3D human motions of the desirable action interacting with the specified objects. For example, sit on the armchair near the desk.To fill in the gap. We collect a large-scale and semantic-rich synthetic human-scene interaction dataset, denoted as HUMANISE, to enable such langauge-conditoned scene understanding tasks.
|
|
EgoTaskQA: Understanding Human Tasks in Egocentric Videos
Baoxiong Jia ,
Ting Lei,
Song-Chun Zhu,
Siyuan Huang
NeurIPS 2022 (Dataset and Benchmark Track)
Paper / Project / Code
we introduce the EgoTaskQA benchmark that8
provides a single home for the crucial dimensions of task understanding through
question answering on real-world egocentric videos. We meticulously design
questions that target the understanding of (1) action dependencies and effects,
(2) intents and goals, and (3) agents' beliefs about others.
|
|
Learning V1 simple cells with vector representations of local contents and matrix representations of local motions
Ruiqi Gao ,
Jianwen Xie ,
Siyuan Huang,
Yufan Ren,
Song-Chun Zhu,
Ying Nian Wu
AAAI 2022
Paper
we propose a representational model that couples the vector representations of local image contents with the matrix representations of local pixel displacements. When the image changes from one time frame to the next due to pixel displacements, the vector at each pixel is rotated by a matrix that represents the displacement of this pixel.
|
|
Spatio-temporal Self-Supervised Representation Learning for 3D Point Clouds
Siyuan Huang*,
Yichen Xie*
Song-Chun Zhu,
Yixin Zhu
ICCV 2021
Paper /
Supplementary /
Project /
Code
We introduce a spatio-temporal representation
learning (STRL) framework, capable of learning from unlabeled 3D point clouds in a self-supervised fashion.
|
|
Inter-GPS: Interpretable Geometry Problem Solving with Formal Language and Symbolic Reasoning
Pan Lu*, Ran Gong*, Shibiao Jiang*, Liang Qiu, Siyuan Huang, Xiaodan Liang, Song-Chun Zhu
ACL 2021
Oral Presentation
Paper /
Code /
Project /
Bibtex
we construct a new largescale benchmarkconsisting of 3,002 geometry problems with dense annotation in formal language and propose a novel geometry solving approach with formal language and symbolic reasoning.
|
|
Learning Neural Representation of Camera Pose with Matrix Representation of Pose Shift via View Synthesis
Yaxuan Zhu,
Ruiqi Gao,
Siyuan Huang,
Song-Chun Zhu,
Ying Nian Wu
CVPR 2021
Oral Presentation
Paper /
Supplementary /
Code
To efficiently represent camera pose in 3D computer vision, we propose an approach to learn neural representations of camera poses and 3D scenes, coupled with neural representations of local camera movements.
|
|
A Competence-aware Curriculum for Visual Concepts Learning via Question Answering
Qing Li ,
Siyuan Huang,
Yining Hong,
Song-Chun Zhu
ECCV 2020
Oral Presentation
Paper
We design a neural-symbolic concept learner for learning the visual concepts and a multi-dimensional Item Response Theory (mIRT) model for guiding the visual concept learning process with an adaptive curriculum.
|
|
LEMMA: A Multi-view Dataset for Learning Multi-agent Multi-task Activities
Baoxiong Jia,
Yixin Chen,
Siyuan Huang,
Yixin Zhu,
Song-Chun Zhu
ECCV 2020
Paper /
Code /
Project /
Bibtex
We introduce the LEMMA dataset to provide a single home to address these missing dimensions with carefully designed settings, wherein the numbers of tasks and agents vary to highlight different learning objectives. We densely annotate the atomic-actions with human-object interactions to provide ground-truth of the compositionality, scheduling, and assignment of daily activities.
|
|
Dark, Beyond Deep: A Paradigm Shift to Cognitive AI with Humanlike Common Sense
Yixin Zhu ,
Tao Gao ,
Lifeng Fan ,
Siyuan Huang,
Edmonds Mark,
Hangxin Liu,
Feng Gao,
Chi Zhang,
Siyuan Qi,
Ying Nian Wu,
Josh Tenenbaum,
Song-Chun Zhu
Engineering 2020
Paper
We demonstrate the power of this perspective to develop cognitive AI systems with humanlike common sense by showing how to
observe and apply FPICU with little training data to solve a wide range of challenging tasks, including tool use, planning, utility
inference, and social learning.
|
|
A Generalized Earley Parser for Human Activity Parsing and Prediction
Siyuan Qi ,
Baoxiong Jia ,
Siyuan Huang,
Ping Wei,
Song-Chun Zhu
TPAMI 2020
Paper
Propose an algorithm to tackle the task of understanding complex human activities from (partially observed) videos from two important aspects: activity recognition and prediction.
|
|
PerspectiveNet: 3D Object Detection from a Single RGB Image via Perspective Points
Siyuan Huang,
Yixin Chen,
Tao Yuan,
Siyuan Qi,
Yixin Zhu,
Song-Chun Zhu
Neural Information Processing Systems (NeurIPS) 2019
Paper
To solve the problem of 3D object detection, we propose perspective points as a novel intermediate representation, defined as the 2D projections of locally-Manhattan 3D keypoints to locate an object, and they satisfy certain geometric constraints caused by the perspective projection.
|
|
Holistic++ Scene Understanding: Single-view 3D Holistic Scene Parsing and Human Pose Estimation with Human-Object Interaction and Physical Commonsense
Yixin Chen *,
Siyuan Huang *,
Tao Yuan,
Siyuan Qi,
Yixin Zhu,
Song-Chun Zhu
IEEE International Conference on Computer Vision (ICCV) 2019
* Equal contributions
Paper /
Supplementary /
Project
Propose a new 3D holistic++ scene understanding problem, which jointly tackles two tasks from a single-view image: (i) holistic scene parsing and reconstruction- and (ii) 3D human pose estimation. We incorporate the human-object interaction (HOI) and physical commonsense to tackle this problem.
|
|
Understanding Human Gaze Communication by Spatio-Temporal Graph Reasoning
Lifeng Fan *,
Wenguan Wang *,
Siyuan Huang,
Xinyu Tang,
Song-Chun Zhu
IEEE International Conference on Computer Vision (ICCV) 2019
* Equal contributions
Paper
Propose a new problem of understanding human gaze communication in social videos from both atomic-level and event-level, which is significant for studying human social interactions.
|
|
Configurable 3D Scene Synthesis and 2D Image Rendering
with Per-Pixel Ground Truth using Stochastic Grammars
* Equal contributions
Internatianal Journal of Computer Vision (IJCV) 2018
Paper /
Demo
Employ physics-based rendering to synthesize photorealistic RGB images while automatically synthesizing detailed,per-pixel ground truth data, including visible surface depth and normal, object identity and material information, as well as illumination.
|
|
Human-centric Indoor Scene Synthesis using Stochastic Grammar
Siyuan Qi,
Yixin Zhu ,
Siyuan Huang,
Chenfanfu Jiang ,
Song-Chun Zhu
IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2018
Paper /
Project /
Code /
Bibtex
Present a human-centric method to sample and synthesize 3D room layouts and 2D images thereof, for the purpose of obtaining large-scale 2D/3D image data with the perfect per-pixel ground truth.
|
|
Predicting Human Activities Using Stochastic Grammar
Siyuan Qi,
Siyuan Huang,
Ping Wei,
Song-Chun Zhu
IEEE International Conference on Computer Vision (ICCV) 2017
Paper /
Bibtex /
Code
Use a stochastic grammar model to capture the compositional structure of events, integrating human actions, objects, and their affordances for modeling the rich context between human and environment.
|
|
Nonlinear Local Metric Learning for Person Re-identification
Siyuan Huang,
Jiwen Lu,
Jie Zhou,
Anil K. Jain
arXiv 2015
arXiv Paper
Utilize the merits of both local metric learning and deep neural network to exploit the complex nonlinear transformations in the feature space of person re-identification data.
|
|
Building Change Detection Based on 3D reconstruction
Baohua Chen,
Lei Deng,
Yueqi Duan,
Siyuan Huang,
Jie Zhou
IEEE International Conference on Image Processing (ICIP) 2015
Paper /
Bibtex
Propose a change detection framework based on RGB-D map generated by 3D reconstruction which can overcome the large illumination changes .
|
|