Ding, Pengxiang

12 publications

ICCV 2025 CARP: Visuomotor Policy Learning via Coarse-to-Fine Autoregressive Prediction Zhefei Gong, Pengxiang Ding, Shangke Lyu, Siteng Huang, Mingyang Sun, Wei Zhao, Zhaoxin Fan, Donglin Wang
AAAI 2025 Cobra: Extending Mamba to Multi-Modal Large Language Model for Efficient Inference Han Zhao, Min Zhang, Wei Zhao, Pengxiang Ding, Siteng Huang, Donglin Wang
ICLR 2025 GEVRM: Goal-Expressive Video Generation Model for Robust Visual Manipulation Hongyin Zhang, Pengxiang Ding, Shangke Lyu, Ying Peng, Donglin Wang
CoRL 2025 Long-VLA: Unleashing Long-Horizon Capability of Vision Language Action Model for Robot Manipulation Yiguo Fan, Shuanghao Bai, Xinyang Tong, Pengxiang Ding, Yuyang Zhu, Hongchao Lu, Fengqi Dai, Wei Zhao, Yang Liu, Siteng Huang, Zhaoxin Fan, Badong Chen, Donglin Wang
ICML 2025 ReinboT: Amplifying Robot Visual-Language Manipulation with Reinforcement Learning Hongyin Zhang, Zifeng Zhuang, Han Zhao, Pengxiang Ding, Hongchao Lu, Donglin Wang
ICML 2025 Rethinking Latent Redundancy in Behavior Cloning: An Information Bottleneck Approach for Robot Manipulation Shuanghao Bai, Wanqi Zhou, Pengxiang Ding, Wei Zhao, Donglin Wang, Badong Chen
NeurIPS 2025 SSR: Enhancing Depth Perception in Vision-Language Models via Rationale-Guided Spatial Reasoning Yang Liu, Ming Ma, Xiaomin Yu, Pengxiang Ding, Han Zhao, Mingyang Sun, Siteng Huang, Donglin Wang
ICML 2025 Score-Based Diffusion Policy Compatible with Reinforcement Learning via Optimal Transport Mingyang Sun, Pengxiang Ding, Weinan Zhang, Donglin Wang
ICLR 2025 VLAS: Vision-Language-Action Model with Speech Instructions for Customized Robot Manipulation Wei Zhao, Pengxiang Ding, Zhang Min, Zhefei Gong, Shuanghao Bai, Han Zhao, Donglin Wang
AAAI 2024 Expressive Forecasting of 3D Whole-Body Human Motions Pengxiang Ding, Qiongjie Cui, Haofan Wang, Min Zhang, Mengyuan Liu, Donglin Wang
ECCV 2024 PiTe: Pixel-Temporal Alignment for Large Video-Language Model Yang Liu, Pengxiang Ding, Siteng Huang, Min Zhang, Han Zhao, Donglin Wang
ECCV 2024 QUAR-VLA: Vision-Language-Action Model for Quadruped Robots Pengxiang Ding, Han Zhao, Wenjie Zhang, Wenxuan Song, Min Zhang, Siteng Huang, Ningxi Yang, Donglin Wang