TY - JOUR
T1 - Flood detection via twitter streams using textual and visual features
AU - Alam, Firoj
AU - Hassan, Zohaib
AU - Ahmad, Kashif
AU - Gul, Asma
AU - Riegler, Michael Alexander
AU - Conci, Nicola
AU - Al-Fuqaha, Ala
N1 - Publisher Copyright:
© 2020 Copyright 2020 for this paper by its authors. All Rights Reserved.
PY - 2020
Y1 - 2020
N2 - The paper presents our proposed solutions for the MediaEval 2020 Flood-Related Multimedia Task, which aims to analyze and detect flooding events in multimedia content shared over Twitter. In total, we proposed four different solutions including a multi-modal solution combining textual and visual information for the mandatory run, and three single modal image and text-based solutions as optional runs. In the multi-modal method, we rely on a supervised multimodal bitransformer model that combines textual and visual features in an early fusion, achieving a micro F1-score of .859 on the development data set. For the text-based flood events detection, we use a transformer network (i.e., pretrained Italian BERT model) achieving an F1-score of .853. For image-based solutions, we employed multiple deep models, pre-trained on both, the ImageNet and Places data sets, individually and combined in an early fusion achieving F1-scores of .816 and .805 on the development set, respectively.
AB - The paper presents our proposed solutions for the MediaEval 2020 Flood-Related Multimedia Task, which aims to analyze and detect flooding events in multimedia content shared over Twitter. In total, we proposed four different solutions including a multi-modal solution combining textual and visual information for the mandatory run, and three single modal image and text-based solutions as optional runs. In the multi-modal method, we rely on a supervised multimodal bitransformer model that combines textual and visual features in an early fusion, achieving a micro F1-score of .859 on the development data set. For the text-based flood events detection, we use a transformer network (i.e., pretrained Italian BERT model) achieving an F1-score of .853. For image-based solutions, we employed multiple deep models, pre-trained on both, the ImageNet and Places data sets, individually and combined in an early fusion achieving F1-scores of .816 and .805 on the development set, respectively.
UR - http://www.scopus.com/inward/record.url?scp=85108056228&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85108056228
SN - 1613-0073
VL - 2882
JO - CEUR Workshop Proceedings
JF - CEUR Workshop Proceedings
T2 - Multimedia Evaluation Benchmark Workshop 2020, MediaEval 2020
Y2 - 14 December 2020 through 15 December 2020
ER -