|
74 | 74 | </tr> |
75 | 75 | </table> |
76 | 76 |
|
77 | | - <table width="880" border="0" align="center" cellspacing="0" cellpadding="0"> |
78 | | - <tr style="background-color: var(--highlight-color)"> |
79 | | - <td style="width:35%; vertical-align:middle; padding-right: 20px;"> |
80 | | - <div class="image-container"> |
81 | | - <img src='publications/2025_MIP.jpg' width="100%"> |
82 | | - </div> |
83 | | - </td> |
84 | | - <td style="width:65%; vertical-align:middle"> |
85 | | - <papertitle>Much Ado About Noising: Dispelling the Myths of Generative Robotic Control</papertitle> |
86 | | - <br> |
87 | | - Chaoyi Pan, Giri Anantharaman, Nai-Chieh Huang, Claire Jin, Daniel Pfrommer, Chenyang Yuan, Frank Permenter, Guannan Qu<sup>†</sup>, Nicholas Boffi<sup>†</sup>, Guanya Shi<sup>†</sup>, Max Simchowitz<sup>†</sup> |
88 | | - <br> |
89 | | - <a href="https://arxiv.org/abs/2512.01809" target="_blank"><i class="far fa-file"></i> paper</a>   |
90 | | - <a href="https://simchowitzlabpublic.github.io/much-ado-about-noising-project/" target="_blank"><i class="fas fa-globe"></i> website</a>   |
91 | | - <a href="https://github.com/simchowitzlabpublic/much-ado-about-noising" target="_blank"><i class="fas fa-code"></i> code</a> |
92 | | - <p style="margin-top: 5px"><i class="fas fa-comment-dots"></i> TL;DR: In most benchmarks, the success of generative policies is NOT from its distributional-learning formulation. |
93 | | - </td> |
94 | | - </tr> |
95 | | - </table> |
96 | | - |
97 | | - <br> |
98 | | - |
99 | 77 | <table width="880" border="0" align="center" cellspacing="0" cellpadding="0"> |
100 | 78 | <tr> |
101 | 79 | <td style="width:35%; vertical-align:middle; padding-right: 20px;"> |
|
182 | 160 |
|
183 | 161 | <br> |
184 | 162 |
|
185 | | - <table width="880" border="0" align="center" cellspacing="0" cellpadding="0"> |
186 | | - <tr style="background-color: var(--highlight-color)"> |
187 | | - <td style="width:35%; vertical-align:middle; padding-right: 20px;"> |
188 | | - <div class="image-container"> |
189 | | - <img src='publications/2025_BFM.gif' width="90%"> |
190 | | - </div> |
191 | | - </td> |
192 | | - <td style="width:65%; vertical-align:middle"> |
193 | | - <papertitle>BFM-Zero: A Promptable Behavioral Foundation Model for Humanoid Control Using Unsupervised Reinforcement Learning</papertitle> |
194 | | - <br> |
195 | | - Yitang Li<sup>*</sup>, Zhengyi Luo<sup>*</sup>, Tonghe Zhang, Cunxi Dai, Anssi Kanervisto, Andrea Tirinzoni, Haoyang Weng, Kris Kitani, Mateusz Guzek, Ahmed Touati, Alessandro Lazaric, Matteo Pirotta<sup>†</sup>, Guanya Shi<sup>†</sup> |
196 | | - <br> |
197 | | - <a href="https://arxiv.org/abs/2511.04131" target="_blank"><i class="far fa-file"></i> paper</a>   |
198 | | - <a href="https://lecar-lab.github.io/BFM-Zero/" target="_blank"><i class="fas fa-globe"></i> website</a>   |
199 | | - <a href="https://github.com/LeCAR-Lab/BFM-Zero" target="_blank"><i class="fas fa-code"></i> code</a> |
200 | | - <p style="margin-top: 5px"><i class="fas fa-comment-dots"></i> TL;DR: BFM-Zero enables zero-shot goal reaching, tracking, and reward optimization (any reward at test time) from one policy. |
201 | | - </td> |
202 | | - </tr> |
203 | | - </table> |
204 | | - |
205 | | - <br> |
206 | | - |
207 | | - <table width="880" border="0" align="center" cellspacing="0" cellpadding="0"> |
208 | | - <tr style="background-color: var(--highlight-color)"> |
209 | | - <td style="width:35%; vertical-align:middle; padding-right: 20px;"> |
210 | | - <div class="image-container"> |
211 | | - <img src='publications/2025_PLD.gif' width="90%"> |
212 | | - </div> |
213 | | - </td> |
214 | | - <td style="width:65%; vertical-align:middle"> |
215 | | - <papertitle>Self-Improving Vision-Language-Action Models with Data Generation via Residual RL</papertitle> |
216 | | - <br> |
217 | | - Wenli Xiao<sup>*</sup>, Haotian Lin<sup>*</sup>, Andy Peng, Haoru Xue, Tairan He, Yuqi Xie, Fengyuan Hu, Jimmy Wu, Zhengyi Luo, Linxi "Jim" Fan, Guanya Shi, Yuke Zhu |
218 | | - <br> |
219 | | - <a href="https://arxiv.org/abs/2511.00091" target="_blank"><i class="far fa-file"></i> paper</a>   |
220 | | - <a href="https://www.wenlixiao.com/self-improve-VLA-PLD" target="_blank"><i class="fas fa-globe"></i> website</a> |
221 | | - <p style="margin-top: 5px"><i class="fas fa-comment-dots"></i> TL;DR: Probe, Learn, Distill (PLD): On-policy probing from a base VLA model + off-policy residual RL + distillation for VLA post-training. |
222 | | - </td> |
223 | | - </tr> |
224 | | - </table> |
225 | | - |
226 | | - <br> |
227 | | - |
228 | 163 | <table width="880" border="0" align="center" cellspacing="0" cellpadding="0"> |
229 | 164 | <tr> |
230 | 165 | <td style="width:35%; vertical-align:middle; padding-right: 20px;"> |
|
374 | 309 | </tr> |
375 | 310 | </table> |
376 | 311 |
|
| 312 | + <br> |
| 313 | + <br> |
| 314 | + |
| 315 | + <table width="880" border="0" align="center" cellspacing="0" cellpadding="0"> |
| 316 | + <tr> |
| 317 | + <td width="100%" valign="middle"> |
| 318 | + <heading>2026</heading> |
| 319 | + </td> |
| 320 | + </tr> |
| 321 | + </table> |
| 322 | + |
| 323 | + <table width="880" border="0" align="center" cellspacing="0" cellpadding="0"> |
| 324 | + <tr style="background-color: var(--highlight-color)"> |
| 325 | + <td style="width:35%; vertical-align:middle; padding-right: 20px;"> |
| 326 | + <div class="image-container"> |
| 327 | + <img src='publications/2025_MIP.jpg' width="100%"> |
| 328 | + </div> |
| 329 | + </td> |
| 330 | + <td style="width:65%; vertical-align:middle"> |
| 331 | + <papertitle>Much Ado About Noising: Dispelling the Myths of Generative Robotic Control</papertitle> |
| 332 | + <br> |
| 333 | + Chaoyi Pan, Giri Anantharaman, Nai-Chieh Huang, Claire Jin, Daniel Pfrommer, Chenyang Yuan, Frank Permenter, Guannan Qu<sup>†</sup>, Nicholas Boffi<sup>†</sup>, Guanya Shi<sup>†</sup>, Max Simchowitz<sup>†</sup> |
| 334 | + <br> |
| 335 | + <em>International Conference on Learning Representations (ICLR)</em>, 2026 |
| 336 | + <br> |
| 337 | + <a href="https://arxiv.org/abs/2512.01809" target="_blank"><i class="far fa-file"></i> paper</a>   |
| 338 | + <a href="https://simchowitzlabpublic.github.io/much-ado-about-noising-project/" target="_blank"><i class="fas fa-globe"></i> website</a>   |
| 339 | + <a href="https://github.com/simchowitzlabpublic/much-ado-about-noising" target="_blank"><i class="fas fa-code"></i> code</a> |
| 340 | + <p style="margin-top: 5px"><i class="fas fa-comment-dots"></i> TL;DR: In most benchmarks, the success of generative policies is NOT from its distributional-learning formulation. |
| 341 | + </td> |
| 342 | + </tr> |
| 343 | + </table> |
| 344 | + |
| 345 | + <br> |
| 346 | + |
| 347 | + <table width="880" border="0" align="center" cellspacing="0" cellpadding="0"> |
| 348 | + <tr style="background-color: var(--highlight-color)"> |
| 349 | + <td style="width:35%; vertical-align:middle; padding-right: 20px;"> |
| 350 | + <div class="image-container"> |
| 351 | + <img src='publications/2025_BFM.gif' width="90%"> |
| 352 | + </div> |
| 353 | + </td> |
| 354 | + <td style="width:65%; vertical-align:middle"> |
| 355 | + <papertitle>BFM-Zero: A Promptable Behavioral Foundation Model for Humanoid Control Using Unsupervised Reinforcement Learning</papertitle> |
| 356 | + <br> |
| 357 | + Yitang Li<sup>*</sup>, Zhengyi Luo<sup>*</sup>, Tonghe Zhang, Cunxi Dai, Anssi Kanervisto, Andrea Tirinzoni, Haoyang Weng, Kris Kitani, Mateusz Guzek, Ahmed Touati, Alessandro Lazaric, Matteo Pirotta<sup>†</sup>, Guanya Shi<sup>†</sup> |
| 358 | + <br> |
| 359 | + <em>International Conference on Learning Representations (ICLR)</em>, 2026 |
| 360 | + <br> |
| 361 | + <a href="https://arxiv.org/abs/2511.04131" target="_blank"><i class="far fa-file"></i> paper</a>   |
| 362 | + <a href="https://lecar-lab.github.io/BFM-Zero/" target="_blank"><i class="fas fa-globe"></i> website</a>   |
| 363 | + <a href="https://github.com/LeCAR-Lab/BFM-Zero" target="_blank"><i class="fas fa-code"></i> code</a> |
| 364 | + <p style="margin-top: 5px"><i class="fas fa-comment-dots"></i> TL;DR: BFM-Zero enables zero-shot goal reaching, tracking, and reward optimization (any reward at test time) from one policy. |
| 365 | + </td> |
| 366 | + </tr> |
| 367 | + </table> |
| 368 | + |
| 369 | + <br> |
| 370 | + |
| 371 | + <table width="880" border="0" align="center" cellspacing="0" cellpadding="0"> |
| 372 | + <tr style="background-color: var(--highlight-color)"> |
| 373 | + <td style="width:35%; vertical-align:middle; padding-right: 20px;"> |
| 374 | + <div class="image-container"> |
| 375 | + <img src='publications/2025_PLD.gif' width="90%"> |
| 376 | + </div> |
| 377 | + </td> |
| 378 | + <td style="width:65%; vertical-align:middle"> |
| 379 | + <papertitle>Self-Improving Vision-Language-Action Models with Data Generation via Residual RL</papertitle> |
| 380 | + <br> |
| 381 | + Wenli Xiao<sup>*</sup>, Haotian Lin<sup>*</sup>, Andy Peng, Haoru Xue, Tairan He, Yuqi Xie, Fengyuan Hu, Jimmy Wu, Zhengyi Luo, Linxi "Jim" Fan, Guanya Shi, Yuke Zhu |
| 382 | + <br> |
| 383 | + <em>International Conference on Learning Representations (ICLR)</em>, 2026 |
| 384 | + <br> |
| 385 | + <a href="https://arxiv.org/abs/2511.00091" target="_blank"><i class="far fa-file"></i> paper</a>   |
| 386 | + <a href="https://www.wenlixiao.com/self-improve-VLA-PLD" target="_blank"><i class="fas fa-globe"></i> website</a> |
| 387 | + <p style="margin-top: 5px"><i class="fas fa-comment-dots"></i> TL;DR: Probe, Learn, Distill (PLD): On-policy probing from a base VLA model + off-policy residual RL + distillation for VLA post-training. |
| 388 | + </td> |
| 389 | + </tr> |
| 390 | + </table> |
| 391 | + |
377 | 392 | <br> |
378 | 393 |
|
379 | 394 | <table width="880" border="0" align="center" cellspacing="0" cellpadding="0"> |
|
388 | 403 | <br> |
389 | 404 | Yuanhang Zhang, Yifu Yuan, Prajwal Gurunath, Tairan He, Shayegan Omidshafiei, Ali-akbar Agha-mohammadi, Marcell Vazquez-Chanlatte, Liam Pedersen, Guanya Shi |
390 | 405 | <br> |
| 406 | + <em>Learning for Dynamics and Control Conference (L4DC)</em>, 2026 |
| 407 | + <br> |
| 408 | + <p style="color: orange; margin: 0px 0;">(Oral Presentation)</p> |
391 | 409 | <a href="https://arxiv.org/abs/2505.06776" target="_blank"><i class="far fa-file"></i> paper</a>   |
392 | 410 | <a href="https://lecar-lab.github.io/falcon-humanoid/" target="_blank"><i class="fas fa-globe"></i> website</a>   |
393 | 411 | <a href="https://github.com/LeCAR-Lab/FALCON/" target="_blank"><i class="fas fa-code"></i> code</a> |
|
399 | 417 | <br> |
400 | 418 |
|
401 | 419 | <table width="880" border="0" align="center" cellspacing="0" cellpadding="0"> |
402 | | - <tr style="background-color: var(--highlight-color)"> |
| 420 | + <tr> |
403 | 421 | <td style="width:35%; vertical-align:middle; padding-right: 20px;"> |
404 | 422 | <div class="image-container"> |
405 | 423 | <img src='publications/2025_TDMPC_Square.gif' width="95%"> |
|
410 | 428 | <br> |
411 | 429 | Haotian Lin, Pengcheng Wang, Jeff Schneider, Guanya Shi |
412 | 430 | <br> |
| 431 | + <em>Learning for Dynamics and Control Conference (L4DC)</em>, 2026 |
| 432 | + <br> |
413 | 433 | <a href="https://arxiv.org/abs/2502.03550v1" target="_blank"><i class="far fa-file"></i> paper</a>   |
414 | 434 | <a href="https://darthutopian.github.io/tdmpc_square/" target="_blank"><i class="fas fa-globe"></i> website</a>   |
415 | 435 | <a href="https://github.com/DarthUtopian/tdmpc_square_public" target="_blank"><i class="fas fa-code"></i> code</a> |
|
418 | 438 | </tr> |
419 | 439 | </table> |
420 | 440 |
|
| 441 | + |
421 | 442 | <br> |
422 | 443 | <br> |
423 | 444 |
|
|
0 commit comments