I am a first-year PhD student in a combined master’s–PhD program in Software Engineering at the University of Electronic Science and Technology of China (UESTC), under the supervision of Prof Fan Zhou. Previously, I received Bachelor of Engineering degree from Fuzhou University.

My research mainly focuses on Robust & Personalized Multimodal Intelligence in real-world, non-ideal, and dynamic conditions. Specifically, I am enthusiastic about designing multimodal systems that can perform effectively under (1) imperfect inputs and environments (e.g., modality missing, distribution shifts, weak supervision) and (2) user-specific dynamics (e.g., MLLM personalization). Moreover, I am also interested in multimodal video understanding and detection, where I apply robustness techniques to improve the generalization, reliability, and robustness of detection models in real-world scenarios.

Feel free to contact me if you have any questions about my research or potential collaboration opportunities.

🔥 News

  • 2026.05:  🎉🎉 1 Paper is accepted by ICML 2026! See you in Seoul!
  • 2026.04:  💦💦 We release the first comprehensive repository of resources on modality-missing learning at awesome-modality-missing-learning .
  • 2026.04:  🎉🎉 1 Paper is accepted by ACL 2026 Findings!
  • 2026.03:  💦💦 2 Papers are submitted to ECCV 2026! The Ship of Theseus now sails again.
  • 2026.02:  🎉🎉 1 Paper is accepted by TCSVT 2026.
  • 2026.02:  🎉🎉 1 Paper is accepted by CVPR 2026 Findings.
  • 2026.02:  💦💦 1 Paper is submitted to KDD 2026 Round 2. Hope a wonderful result.
  • 2025.11:  🎉🎉 3 Papers are accepted by KDD 2026! See you in Jeju!
  • 2025.10:  🎉🎉 Get Postgraduate National Scholarship again.

📝 Selected Publications (*=Equal Contribution, †=Conresponding Author)

🛡 Robust Multimodal Learning

🧩 Robust Against Missing Modalities

ICML 2026
sym

AOEPT: Breaking the Implicit Modality-Reduction Bottleneck in Modality Missing Prompt Tuning

Jian Lang, Rongpei Hong, Ting Zhong, Fan Zhou†

ICML 2026 | CCF A | PDF | Github

  • The Implicit Modality-Reduction (IMR) bottleneck in existing modality-missing prompt-tuning methods, and a new metric Normalized Missing-modality Mutual Information (NM2I) quantifies IMR.
  • A minimalist Modal-Contextualized Prompting method (AOEPT) breaks IMR.
KDD 2025
sym

REDEEMing Modality Information Loss: Retrieval-Guided Conditional Generation for Severely Modality Missing Learning

Jian Lang, Rongpei Hong, Zhangtao Cheng, Ting Zhong, Fan Zhou†

KDD 2025 | CCF A | PDF | Github

  • REDEEM, the extension work of our RAGPT.
  • Proposing a retrieval-guided conditional generation paradigm for enhancing the modality-missing robustness of pre-trained Multimodal Transformer.
AAAI 2025
sym

Retrieval-Augmented Dynamic Prompt Tuning for Incomplete Multimodal Learning

Jian Lang*, Zhangtao Cheng*, Ting Zhong, Fan Zhou†

AAAI 2025 | CCF A | PDF | Github |

  • RAGPT, a novel retrieval-augmented dynamic prompt-tuning framework for enhancing the modality-missing robustness of pre-trained Multimodal Transformer.

⚓ Robust Against Domain (Distribution) Shift

KDD 2026
sym

Nip Rumors in the Bud: Retrieval-Guided Topic-Level Adaptation for Test-Time Fake News Video Detection

Jian Lang, Rongpei Hong, Ting Zhong, Yong Wang, Fan Zhou†

KDD 2026 | CCF A | PDF | Github

  • RADAR, the first work to achieves the test-time adaptation of the Fake News Video Detection,
  • Enabling fast adaptation to evolving news videos with shifting topic-level distributions in the dynamic world.
AAAI 2026
sym

Shedding the Facades, Connecting the Domains: Detecting Shifting Multimodal Hate Video with Test-Time Adaptation

Jiao Li, Jian Lang, Xikai Tang†, Ting Zhong, Fan Zhou

AAAI 2026 | CCF A | PDF | Github

  • SCANNER, the first test-time adaptation framework tailored for distribution shifting hate video detection.

🪙 Robust Against Data / Label Scarcity

KDD 2026
sym

From Shallow Humor to Metaphor: Towards Label-Free Harmful Meme Detection via LMM Agent Self-Improvement

Jian Lang, Rongpei Hong, Ting Zhong, Leiting Chen, Qiang Gao, Fan Zhou†

KDD 2026 | CCF A | PDF | Github

  • ALARM, the first label-free harmful meme detection framework powered by LMM self-improvement
  • Enabling prompt and robust adaptation to evolving topics and themes of harmful web memes.
ICCV 2025
sym

Borrowing Eyes for the Blind Spot: Overcoming Data Scarcity in Malicious Video Detection via Cross-Domain Retrieval Augmentation

Rongpei Hong*, Jian Lang*, Ting Zhong, Fan Zhou†

ICCV 2025 | CCF A | PDF | Github

  • CRAVE, a novel cross-domain retrieval augmentation framework that transfers knowledge from resource-rich image-text domain to enhance malicious video detection with scarce training data.

🧑‍🦱 Personalized Multimodal Learning

🧏 MLLM Personalized Understanding

KDD 2026
sym

TAMEing Long Contexts in Personalization: Towards Training-Free and State-Aware MLLM Personalized Assistant

Rongpei Hong, Jian Lang, Ting Zhong†, Yong Wang, Fan Zhou

KDD 2026 | CCF A | PDF | Github

  • TAME, the first training-free and state-aware personalized Multimodal Large Multimodal Model assistant powered by double memories.

🎥 Video Analysis & Detection

ACL 2026
sym

LEAF: Towards Lightweight Explainable Hateful Video Detection via Self-Grounding CoT Guided Stage-Wise Distillation

Jian Lang, Rongpei Hong, Meihui Zhong, Kaiju Li, Ting Zhong, Qiang Gao, Fan Zhou†

ACL 2026 Findings | PDF | Github

  • LEAF, the first lightweight while explainable hateful video detection framework powered by SG-CoT guided MLLM distillation.
TCSVT 2026
sym

MATCH: Multi-Agentic Evidence Grounding for Explainable Hate Video Detection

Kaiju Li, Rongpei Hong, Jian Lang, Jin Wu†, Fan Zhou†, Jingkuan Song

TCSVT 2026 | CAS Q1 Top | PDF

  • MATCH, a novel multiple LMM agent collaboration framework for interpretable hate video detection.
WWW 2025
sym

Biting Off More Than You Can Detect: Retrieval-Augmented Multimodal Experts for Short Video Hate Detection

Jian Lang, Rongpei Hong, Jin Xu, Xovee Xu, Yili Li, Fan Zhou†

WWW 2025 | CCF A | PDF | Github

  • MoRE, a novel mixture of retrieval-augmented multimodal experts framework to enhance hate video detection.
WWW 2025
sym

Following Clues, Approaching the Truth: Explainable Micro-Video Rumor Detection via Chain-of-Thought Reasoning

Rongpei Hong, Jian Lang, Jin Xu, Zhangtao Cheng, Ting Zhong†, Fan Zhou

WWW 2025 | CCF A | PDF | Github

  • ExMRD, the first explainable fake news video detection framework powered by the Chain-of-Thought Reasoning.
ICME 2025
sym

REAL: Retrieval-Augmented Prototype Alignment for Improved Fake News Video Detection

Yili Li, Jian Lang, Rongpei Hong, Qing Chen, Zhangtao Cheng, Jia Chen, Ting Zhong, Fan Zhou†

ICME 2025 | CCF B | PDF | Github |

  • REAL, a novel model-agnostic framework that generates manipulation-aware representations to enhance existing methods in detecting fake news videos with only subtle modifications to the original authentic ones.
SIGIR 2024
sym

Predicting Micro-video Popularity via Multi-modal Retrieval Augmentation

Ting Zhong, Jian Lang, Yifan Zhang, Zhangtao Cheng, Kunpeng Zhang, Fan Zhou†

SIGIR 2024 | CCF A | PDF | Github |

  • MMRA, a multi-modal retrieval-augmented popularity prediction model that enhances prediction accuracy using relevant retrieved information.

🎖 Honors and Awards

  • 2025.10 National Scholarship (Top 1%)
  • 2025.10 Master’s Student Academic Scholarship (1st Division, Ranked 1st)
  • 2024.10 National Scholarship (Top 1%)
  • 2024.10 Master’s Student Academic Scholarship (1st Division, Ranked 1st)
  • 2023.12 Artificial Intelligence Algorithm Challenge Runner-up (2nd), hosted by People’s Daily Online

📖 Educations

  • 2023.09 -, PhD Student, University of Electronic Science and Technology of China
  • 2019.09 - 2023.06, Undergraduate, Fuzhou University

📝 Peer Review

  • Conference Review: NeurIPS 2026 Reviewer, KDD 2026 Reviewer, ICML 2026 (Emergency) Reviewer, AAAI 2026 Reviewer
  • Journal Review: IJCV Reviewer, TPAMI Reviewer, TCSVT Reviewer, KBS Reviewer, ESWA Reviewer

💻 Internships