PulseAugur
LIVE 00:48:19
research · [1 source] ·
0
research

MLLMs struggle with Chinese short-video misinformation, Gemini-2.5-Pro leads

Researchers have developed a new framework to evaluate how well Multimodal Large Language Models (MLLMs) can identify misinformation in Chinese short videos. The study utilized a dataset of 200 videos annotated for deceptive patterns like experimental errors and logical fallacies. Results showed that Gemini-2.5-Pro performed best, achieving a belief score of 71.5, while another model, o3, performed poorly with a score of 35.2. The evaluation also revealed that MLLMs are susceptible to biases, such as those presented by authoritative channel IDs. AI

Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →

IMPACT This research highlights MLLM vulnerabilities to misinformation and biases, suggesting a need for improved robustness in multimodal AI systems.

RANK_REASON This is a research paper introducing a new evaluation framework and dataset for MLLMs.

Read on arXiv cs.CL →

COVERAGE [1]

  1. arXiv cs.CL TIER_1 · Jen-tse Huang, Chang Chen, Shiyang Lai, Wenxuan Wang, Michelle R. Kaufman, Mark Dredze ·

    Probing Multimodal Large Language Models on Cognitive Biases in Chinese Short-Video Misinformation

    arXiv:2601.06600v2 Announce Type: replace Abstract: Short-video platforms have become major channels for misinformation, where deceptive claims frequently leverage visual experiments and social cues. While Multimodal Large Language Models (MLLMs) have demonstrated impressive reas…