Skip to content

Commit 338fb72

Browse files
author
Jiachen Li
committed
Update research
1 parent 98c01df commit 338fb72

File tree

6 files changed

+52
-30
lines changed

6 files changed

+52
-30
lines changed

_data/research.yaml

+51-29
Original file line numberDiff line numberDiff line change
@@ -1,59 +1,38 @@
1-
- title: Explainable Relational Reasoning and Multi-Agent Interaction Modeling (Social & Physical)
1+
- title: Safe and Robust Interaction-Aware Decision Making for Human-Robot Interactions
22
# subtitle: a subtitle
33
# group: featured
4-
image: research_images/Explainable Relational Reasoning.webp
5-
# link: https://github.com/
6-
description: |
7-
We investigate relational reasoning and interaction modeling in the context of the trajectory prediction task, which aims to generate accurate, diverse future trajectory hypotheses or state sequences based on historical observations.
8-
**Our research introduced the first unified relational reasoning toolbox that systematically infers the underlying relations/interactions between entities at different scales (e.g., pairwise, group-wise) and different abstraction levels (e.g., multiplex) by learning dynamic latent interaction graphs and hypergraphs from observable states (e.g., positions) in an unsupervised manner.**
9-
The learned latent graphs are explainable and generalizable, significantly improving the performance of downstream tasks, including prediction, sequential decision making, and control.
10-
**We also proposed a physics-guided relational learning approach for physical dynamics modeling.**<br>
11-
12-
**Related Publications\:** <br>
13-
1. [EvolveGraph: Multi-Agent Trajectory Prediction with Dynamic Relational Reasoning](https://proceedings.neurips.cc/paper/2020/hash/e4d8163c7a068b65a64c89bd745ec360-Abstract.html), NeurIPS 2020. <br>
14-
2. [RAIN: Reinforced Hybrid Attention Inference Network for Motion Forecasting](https://openaccess.thecvf.com/content/ICCV2021/html/Li_RAIN_Reinforced_Hybrid_Attention_Inference_Network_for_Motion_Forecasting_ICCV_2021_paper.html), ICCV 2021. <br>
15-
3. [Interaction Modeling with Multiplex Attention](https://proceedings.neurips.cc/paper_files/paper/2022/hash/7e6361a5d73a8fab093dd8453e0b106f-Abstract-Conference.html), NeurIPS 2022. <br>
16-
4. [Learning Physical Dynamics with Subequivariant Graph Neural Networks](https://proceedings.neurips.cc/paper_files/paper/2022/hash/a845fdc3f87751710218718adb634fe7-Abstract-Conference.html), NeurIPS 2022. <br>
17-
5. [Grouptron: Dynamic Multi-Scale Graph Convolutional Networks for Group-Aware Dense Crowd Trajectory Forecasting](https://arxiv.org/abs/2109.14128), ICRA 2022. <br>
18-
6. [Multi-Agent Dynamic Relational Reasoning for Social Robot Navigation](https://arxiv.org/abs/2401.12275), submitted to IEEE Transactions on Robotics (T-RO), under review.
19-
# repo: greenelab/lab-website-template
20-
tags:
21-
- Graph Neural Networks
22-
23-
- title: Interaction-Aware Decision Making and Model-Based Control
24-
# subtitle: a subtitle
25-
# group: featured
26-
image: research_images/Interaction-Aware Decision Making and Model-Based Control.png
4+
image: research_images/human_robot_interaction.png
275
# link: https://github.com/
286
description: |
297
Although autonomous navigation in simple, static environments has been well studied, it remains challenging for robots to navigate in highly dynamic, interactive scenarios (e.g., intersections, narrow corridors) where humans are involved.
308
Robots must learn a safe and efficient behavior policy that can model the interactions, coordinate with surrounding static/dynamic entities, and generalize to out-of-distribution (OOD) situations.
319
**Our research introduced a novel interaction-aware decision making framework for autonomous vehicles based on reinforcement learning, which integrates human internal state inference, domain knowledge, trajectory prediction, and counterfactual reasoning systematically.
3210
We also investigate model-based control methods that leverage the learned pairwise and group-wise relations for social robot navigation around human crowds.**
33-
Both methods achieve superior performance in the corresponding tasks in terms of a wide range of evaluation metrics and provide explainable, human-understandable intermediate representations to build both users’ and developers’ trust.<br>
11+
Both methods achieve superior performance in the corresponding tasks in terms of a wide range of evaluation metrics and provide explainable, human-understandable intermediate representations to build both users’ and developers’ trust.<br><br>
3412
3513
3614
**Related Publications\:** <br>
37-
1. [Reinforcement Learning for Autonomous Driving with Latent State Inference and Spatial-Temporal Relationships](https://arxiv.org/abs/2011.04251), ICRA 2021. <br>
15+
1. [Multi-Agent Dynamic Relational Reasoning for Social Robot Navigation](https://arxiv.org/abs/2401.12275), submitted to IEEE Transactions on Robotics (T-RO), under review. <br>
16+
2. [Reinforcement Learning for Autonomous Driving with Latent State Inference and Spatial-Temporal Relationships](https://arxiv.org/abs/2011.04251), ICRA 2021. <br>
3817
2. [Autonomous Driving Strategies at Intersections: Scenarios, State-of-the-Art, and Future Outlooks](https://arxiv.org/abs/2106.13052), ITSC 2021. <br>
3918
3. [Robust Driving Policy Learning with Guided Meta Reinforcement Learning](https://arxiv.org/abs/2307.10160), ITSC 2023. <br>
4019
4. [Game Theory-Based Simultaneous Prediction and Planning for Autonomous Vehicle Navigation in Crowded Environments](https://www.researchgate.net/publication/374831905_Game_Theory-Based_Simultaneous_Prediction_and_Planning_for_Autonomous_Vehicle_Navigation_in_Crowded_Environments), ITSC 2023. <br>
4120
5. [Interactive Autonomous Navigation with Internal State Inference and Interactivity Estimation](https://arxiv.org/abs/2311.16091), T-RO 2024. <br>
42-
6. [Multi-Agent Dynamic Relational Reasoning for Social Robot Navigation](https://arxiv.org/abs/2401.12275), submitted to IEEE Transactions on Robotics (T-RO), under review.
4321
# repo: greenelab/lab-website-template
4422
tags:
4523
- Social Navigation
4624

25+
4726
- title: Vision and Language Models for Embodied Intelligence
4827
# subtitle: a subtitle
4928
# group: featured
50-
image: research_images/Vision and Language Models for Embodied Intelligence.png
29+
image: research_images/VLM.png
5130
# link: https://github.com/
5231
description: |
5332
We investigate foundation models and vision language models (VLMs) for robotics and autonomous systems to enhance their reasoning capability and reliability.
5433
For example, inferring the short-term and long-term intentions of traffic participants and understanding the contextual semantics of scenes are the keys to scene understanding and situational awareness of autonomous vehicles.
5534
Moreover, how to enable autonomous agents (e.g., self-driving cars) to explain their reasoning, prediction, and decision making processes to human users (e.g., drivers, passengers) in a human understandable form (e.g., natural language) to build humans’ trust remains largely underexplored.
56-
Therefore, **we created the first multimodal dataset for a new risk object ranking and natural language explanation task in urban scenarios and a rich dataset for intention prediction in autonomous driving, establishing benchmarks for corresponding tasks. Meanwhile, our research introduced novel methods that achieve superior performance on these problems.**<br>
35+
Therefore, **we created the first multimodal dataset for a new risk object ranking and natural language explanation task in urban scenarios and a rich dataset for intention prediction in autonomous driving, establishing benchmarks for corresponding tasks. Meanwhile, our research introduced novel methods that achieve superior performance on these problems.**<br><br>
5736
5837
**Related Publications\:** <br>
5938
1. [LOKI: Long Term and Key Intentions for Trajectory Prediction](https://openaccess.thecvf.com/content/ICCV2021/html/Girase_LOKI_Long_Term_and_Key_Intentions_for_Trajectory_Prediction_ICCV_2021_paper.html), ICCV 2021. <br>
@@ -63,6 +42,49 @@
6342
# repo: greenelab/lab-website-template
6443
tags:
6544
- Foundation Models
45+
- Mobile Manipulation
46+
- Autonomous Driving
47+
- Human-Robot Interaction/Collaboration
48+
- Multi-Agent/Robot Systems
49+
- Cooperative Planning
50+
51+
52+
- title: Explainable and Generalizable Relational Reasoning and Multi-Agent Interaction Modeling (Social & Physical)
53+
# subtitle: a subtitle
54+
# group: featured
55+
image: research_images/Explainable Relational Reasoning.webp
56+
# link: https://github.com/
57+
description: |
58+
We investigate dynamic relational reasoning and interaction modeling under the context of the trajectory/motion prediction task, which aims to generate accurate, diverse future trajectory hypotheses or state sequences based on historical observations.
59+
**Our research introduced the first unified relational reasoning toolbox that systematically infers the underlying relations/interactions between entities at different scales (e.g., pairwise, group-wise) and different abstraction levels (e.g., multiplex) by learning dynamic latent interaction graphs and hypergraphs from observable states (e.g., positions) in an unsupervised manner.**
60+
The learned latent graphs are explainable and generalizable, significantly improving the performance of downstream tasks, including more accurate and generalizable prediction as well as safer and more efficient sequential decision making and control for mobile robots.
61+
**We also proposed a physics-guided relational learning approach for physical dynamics modeling.**<br><br>
62+
63+
**Related Publications\:** <br>
64+
1. [Multi-Agent Dynamic Relational Reasoning for Social Robot Navigation](https://arxiv.org/abs/2401.12275), submitted to IEEE Transactions on Robotics (T-RO). <br>
65+
2. [Grouptron: Dynamic Multi-Scale Graph Convolutional Networks for Group-Aware Crowd Trajectory Forecasting](https://arxiv.org/abs/2109.14128), ICRA 2022. <br>
66+
3. [Learning Physical Dynamics with Subequivariant Graph Neural Networks](https://proceedings.neurips.cc/paper_files/paper/2022/hash/a845fdc3f87751710218718adb634fe7-Abstract-Conference.html), NeurIPS 2022. <br>
67+
4. [Interaction Modeling with Multiplex Attention](https://proceedings.neurips.cc/paper_files/paper/2022/hash/7e6361a5d73a8fab093dd8453e0b106f-Abstract-Conference.html), NeurIPS 2022. <br>
68+
5. [RAIN: Reinforced Hybrid Attention Inference Network for Motion Forecasting](https://openaccess.thecvf.com/content/ICCV2021/html/Li_RAIN_Reinforced_Hybrid_Attention_Inference_Network_for_Motion_Forecasting_ICCV_2021_paper.html), ICCV 2021. <br>
69+
6. [EvolveGraph: Multi-Agent Trajectory Prediction with Dynamic Relational Reasoning](https://proceedings.neurips.cc/paper/2020/hash/e4d8163c7a068b65a64c89bd745ec360-Abstract.html), NeurIPS 2020. <br>
70+
71+
72+
# repo: greenelab/lab-website-template
73+
tags:
74+
- Relational Reasoning
75+
- Graph Neural Networks
76+
- Multi-Agent Interaction
77+
- Trajectory Prediction
78+
- Explainability/Interpretability
79+
- Generalization
80+
- Autonomous Driving
81+
- Social Navigation
82+
- Human-Robot Interaction/Collaboration
83+
84+
85+
86+
87+
6688

6789
- title: Improving Generalizability by Learning Context Relations
6890
# subtitle: a subtitle

research/index.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ nav:
77

88
# {% include icon.html icon="fa-solid fa-robot" %}Featured Research
99

10-
The ultimate goal of our research is to build **trustworthy**, **interactive**, and **human-centered** autonomous agents that can perceive, understand, and reason about the physical world; safely interact and collaborate with humans and other agents, and clearly explain their behaviors to build trust with humans so that they can benefit society in daily lives. To achieve this goal, we have been pursuing interdisciplinary research and unifying the techniques and tools from robotics, machine learning, reinforcement learning, explainable AI, control theory, optimization, and computer vision.
10+
The ultimate goal of our research is to build **trustworthy**, **interactive**, and **human-centered** autonomous embodied agents that can perceive, understand, and reason about the physical world; safely interact and collaborate with humans; and efficiently coordinate with other intelligent agents so that they can benefit society in daily lives. To achieve this goal, we have been pursuing interdisciplinary research and unifying the techniques and tools from robotics, trustworthy AI/ML, deep reinforcement learning, control theory, optimization, and computer vision.
1111

1212
<!-- {% include list_research.html component="card" data="projects" filters="group: featured" %} -->
1313
{% include list_research.html component="card_research" data="research"%}

research_images/TIAGo-IX.JPG

1.62 MB
Loading

research_images/VLM.png

7.63 MB
Loading
5.62 MB
Loading
4.55 MB
Loading

0 commit comments

Comments
 (0)