|
Postdoctoral Researcher King Abdullah University of Science and Technology |
|
|
|
|
My research aims to develop AI systems that understand complex visual environments and bridge vision and language. My interests include video understanding, question answering, and visual grounding. Recently, I focus on egocentric video understanding and multimodal large language models for real-world and healthcare applications, with an emphasis on efficiency, interpretability, and trustworthy AI. My goal is to build intelligent systems that understand the physical world, support human decision-making, and enable effective human–AI interaction. I am actively seeking Research Interns/Research Assistants/Visiting Students with backgrounds in CV/NLP/Multimodal. |
|
|
My group is actively looking for fully funded visiting students/interns (free housing, flight ticket, medical insurance, plus 1000 USD/month for stipend). If you are interested, please contact me. |
|
|
|
|
* Equal contribution; # Core Contributor; † Corresponding author |
|
|
Yang Shi#, Yuhao Dong#, Yue Ding#, Yuran Wang#, Xuanyu Zhu#, Sheng Zhou#, Wenting Liu#, Haochen Tian#, Rundong Wang#, Huanqian Wang, Zuyan Liu, Bohan Zeng, Ruizhe Chen, Qixun Wang, Zhuoran Zhang, Xinlong Chen, Chengzhuo Tong, Bozhou Li, Qiang Liu, Haotian Wang†, Wenjing Yang, Yuanxing Zhang†, Pengfei Wan, YiFan Zhang†, Ziwei Liu†. CVPR'26 [arXiv] [Code] [Dataset] |
|
Sheng Zhou, Junbin Xiao†, Qingyun Li, Yicong Li, Xun Yang, Dan Guo, Meng Wang, Tat-Seng Chua, Angela Yao. CVPR'25 [arXiv] [Project Page] [Code] [Dataset] |
|
Sheng Zhou, Junbin Xiao†, Xun Yang†, Peipei Song, Dan Guo†, Angela Yao, Meng Wang, Tat-Seng Chua. IEEE TMM'25 [arXiv] [Code] [Dataset] |
|
Sheng Zhou, Dan Guo†, Xun Yang†, Jianfeng Dong, Meng Wang†. ACM TOMM'24 [Paper] [Code] |
|
Sheng Zhou, Dan Guo†, Jia Li, Xun Yang†, Meng Wang†. IEEE TIP'23 [Paper] [Code] |
|
|
|
|
|
|