论文标题

社交媒体中危机事件的多模式分类

Multimodal Categorization of Crisis Events in Social Media

论文作者

Abavisani, Mahdi, Wu, Liwei, Hu, Shengli, Tetreault, Joel, Jaimes, Alejandro

论文摘要

图像分类和自然语言处理方面的最新发展,再加上社交媒体使用的快速增长,在实时检测世界各地的破裂事件方面取得了基本进步。紧急响应是从这些进步中获得的这样的领域。通过每分钟处理数十亿个文本和图像,可以自动检测到事件,以使应急响应工作者能够更好地评估迅速发展的情况并相应地部署资源。迄今为止,该领域的大多数事件检测技术都集中在仅图像或仅文本方法上,从而限制了检测性能并影响传递给危机响应团队的信息质量。在本文中,我们提出了一种新的多模式融合方法,该方法利用图像和文本作为输入。特别是,我们引入了一个交叉意见模块,该模块可以按样本按样本中的弱模态过滤弱和误导性的组件。此外,我们在训练过程中采用了一种基于多模式的基于图形的方法,在不同多模式对的嵌入之间进行随机过渡,从而更好地规范学习过程,并通过构建来自不同样本的新匹配对来处理有限的训练数据。我们表明,我们的方法在三个与危机相关的任务上的优势优于单峰方法和强大的多模式基线。

Recent developments in image classification and natural language processing, coupled with the rapid growth in social media usage, have enabled fundamental advances in detecting breaking events around the world in real-time. Emergency response is one such area that stands to gain from these advances. By processing billions of texts and images a minute, events can be automatically detected to enable emergency response workers to better assess rapidly evolving situations and deploy resources accordingly. To date, most event detection techniques in this area have focused on image-only or text-only approaches, limiting detection performance and impacting the quality of information delivered to crisis response teams. In this paper, we present a new multimodal fusion method that leverages both images and texts as input. In particular, we introduce a cross-attention module that can filter uninformative and misleading components from weak modalities on a sample by sample basis. In addition, we employ a multimodal graph-based approach to stochastically transition between embeddings of different multimodal pairs during training to better regularize the learning process as well as dealing with limited training data by constructing new matched pairs from different samples. We show that our method outperforms the unimodal approaches and strong multimodal baselines by a large margin on three crisis-related tasks.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源