TY - JOUR
T1 - Floods detection in twitter text and images
AU - Said, Naina
AU - Ahmad, Kashif
AU - Gul, Asma
AU - Ahmad, Nasir
AU - Al-Fuqaha, Ala
N1 - Publisher Copyright:
© 2020 Copyright 2020 for this paper by its authors. All Rights Reserved.
PY - 2020
Y1 - 2020
N2 - In this paper, we present our methods for the MediaEval 2020 Flood Related Multimedia task, which aims to analyze and combine textual and visual content from social media for the detection of real-world flooding events. The task mainly focuses on identifying floods related tweets relevant to a specific area. We propose several schemes to address the challenge. For text-based flood events detection, we use three different methods, relying on Bag of Words (BOW) and an Italian Version of Bert individually and in combination, achieving an F1-score of 0.77%, 0.68%, and 0.70% on the development set, respectively. For the visual analysis, we rely on features extracted via multiple state-of-the-art deep models pre-trained on ImageNet. The extracted features are then used to train multiple individual classifiers whose scores are then combined in a late fusion manner achieving an F1-score of 0.75%. For our mandatory multi-modal run, we combine the classification scores obtained with the best textual and visual schemes in a late fusion manner. Overall, better results are obtained with the multimodal scheme achieving an F1-score of 0.80% on the development set.
AB - In this paper, we present our methods for the MediaEval 2020 Flood Related Multimedia task, which aims to analyze and combine textual and visual content from social media for the detection of real-world flooding events. The task mainly focuses on identifying floods related tweets relevant to a specific area. We propose several schemes to address the challenge. For text-based flood events detection, we use three different methods, relying on Bag of Words (BOW) and an Italian Version of Bert individually and in combination, achieving an F1-score of 0.77%, 0.68%, and 0.70% on the development set, respectively. For the visual analysis, we rely on features extracted via multiple state-of-the-art deep models pre-trained on ImageNet. The extracted features are then used to train multiple individual classifiers whose scores are then combined in a late fusion manner achieving an F1-score of 0.75%. For our mandatory multi-modal run, we combine the classification scores obtained with the best textual and visual schemes in a late fusion manner. Overall, better results are obtained with the multimodal scheme achieving an F1-score of 0.80% on the development set.
UR - http://www.scopus.com/inward/record.url?scp=85107586037&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85107586037
SN - 1613-0073
VL - 2882
JO - CEUR Workshop Proceedings
JF - CEUR Workshop Proceedings
T2 - Multimedia Evaluation Benchmark Workshop 2020, MediaEval 2020
Y2 - 14 December 2020 through 15 December 2020
ER -