Follow
Yao (Mark) Mu
Yao (Mark) Mu
The University of Hong Kong, National University of Singapore
Verified email at connect.hku.hk - Homepage
Title
Cited by
Cited by
Year
Embodiedgpt: Vision-language pre-training via embodied chain of thought
Y Mu, Q Zhang, M Hu, W Wang, M Ding, J Jin, B Wang, J Dai, Y Qiao, ...
Advances in Neural Information Processing Systems 36, 2024
1052024
Languagempc: Large language models as decision makers for autonomous driving
H Sha, Y Mu, Y Jiang, L Chen, C Xu, P Luo, SE Li, M Tomizuka, W Zhan, ...
arXiv preprint arXiv:2310.03026, 2023
742023
Adaptdiffuser: Diffusion models as adaptive self-evolving planners
Z Liang, Y Mu, M Ding, F Ni, M Tomizuka, P Luo
arXiv preprint arXiv:2302.01877, 2023
562023
Maskplace: Fast chip placement via reinforced visual representation learning
Y Lai, Y Mu, P Luo
Advances in Neural Information Processing Systems 35, 24019-24030, 2022
312022
Scale-equivalent distillation for semi-supervised object detection
Q Guo, Y Mu, J Chen, T Wang, Y Yu, P Luo
Proceedings of the IEEE/CVF conference on computer vision and pattern …, 2022
282022
IGRNet: a deep learning model for non-invasive, real-time diagnosis of prediabetes through electrocardiograms
L Wang, Y Mu, J Zhao, X Wang, H Che
Sensors 20 (9), 2556, 2020
252020
Separated proportional-integral lagrangian for chance constrained reinforcement learning
B Peng, Y Mu, J Duan, Y Guan, SE Li, J Chen
2021 IEEE Intelligent Vehicles Symposium (IV), 193-199, 2021
212021
Don't Touch What Matters: Task-Aware Lipschitz Data Augmentation for Visual Reinforcement Learning
Z Yuan, G Ma, Y Mu, B Xia, B Yuan, X Wang, P Luo, H Xu
arXiv preprint arXiv:2202.09982, 2022
192022
Enhance sample efficiency and robustness of end-to-end urban autonomous driving via semantic masked world model
Z Gao, Y Mu, C Chen, J Duan, P Luo, Y Lu, SE Li
IEEE Transactions on Intelligent Transportation Systems, 2024
182024
Metadiffuser: Diffusion model as conditional planner for offline meta-rl
F Ni, J Hao, Y Mu, Y Yuan, Y Zheng, B Wang, Z Liang
International Conference on Machine Learning, 26087-26105, 2023
182023
Model-based chance-constrained reinforcement learning via separated proportional-integral lagrangian
B Peng, J Duan, J Chen, SE Li, G Xie, C Zhang, Y Guan, Y Mu, E Sun
IEEE Transactions on Neural Networks and Learning Systems 35 (1), 466-478, 2022
182022
Tree-planner: Efficient close-loop task planning with large language models
M Hu, Y Mu, X Yu, M Ding, S Wu, W Shao, Q Chen, B Wang, Y Qiao, ...
arXiv preprint arXiv:2310.08582, 2023
162023
CO^ 3: Cooperative Unsupervised 3D Representation Learning for Autonomous Driving
R Chen, Y Mu, R Xu, W Shao, C Jiang, H Xu, Z Li, P Luo
arXiv preprint arXiv:2206.04028, 2022
162022
Flow-based recurrent belief state learning for pomdps
X Chen, YM Mu, P Luo, S Li, J Chen
International Conference on Machine Learning, 3444-3468, 2022
152022
Mixed reinforcement learning for efficient policy optimization in stochastic environments
Y Mu, B Peng, Z Gu, SE Li, C Liu, B Nie, J Zheng, B Zhang
2020 20th International Conference on Control, Automation and Systems (ICCAS …, 2020
152020
Ctrlformer: Learning transferable state representation for visual control via transformer
Y Mu, S Chen, M Ding, J Chen, R Chen, P Luo
arXiv preprint arXiv:2206.08883, 2022
142022
Aligndiff: Aligning diverse human preferences via behavior-customisable diffusion model
Z Dong, Y Yuan, J Hao, F Ni, Y Mu, Y Zheng, Y Hu, T Lv, C Fan, Z Hu
arXiv preprint arXiv:2310.02054, 2023
112023
Model-based actor-critic with chance constraint for stochastic system
B Peng, Y Mu, Y Guan, SE Li, Y Yin, J Chen
2021 60th IEEE Conference on Decision and Control (CDC), 4694-4700, 2021
112021
Ec2: Emergent communication for embodied control
Y Mu, S Yao, M Ding, P Luo, C Gan
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2023
102023
Euclid: Towards efficient unsupervised reinforcement learning with multi-choice dynamics model
Y Yuan, J Hao, F Ni, Y Mu, Y Zheng, Y Hu, J Liu, Y Chen, C Fan
arXiv preprint arXiv:2210.00498, 2022
102022
The system can't perform the operation now. Try again later.
Articles 1–20