TY - JOUR
T1 - NLP Techniques for Water Quality Analysis in Social Media Content
AU - Ayub, Muhammad Asif
AU - Ahmad, Khubaib
AU - Ahmad, Kashif
AU - Ahmad, Nasir
AU - Al-Fuqaha, Ala
N1 - Publisher Copyright:
Copyright 2021 for this paper by its authors.
PY - 2021
Y1 - 2021
N2 - This paper presents our contributions to the MediaEval 2021 task namely”WaterMM: Water Quality in Social Multimedia”. The task aims at analyzing social media posts relevant to water quality with particular focus on the aspects like watercolor, smell, taste, and related illnesses. To this aim, a multimodal dataset containing both textual and visual information along with meta-data is provided. Considering the quality and quantity of available content, we mainly focus on textual information by employing three different models individually and jointly in a late-fusion manner. These models include (i) Bidirectional Encoder Representations from Transformers (BERT), (ii) Robustly Optimized BERT Pre-training Approach (XLM-RoBERTa), and a (iii) custom Long short-term memory (LSTM) model obtaining an overall F1-score of 0.794, 0.717, 0.663 on the official test set, respectively. In the fusion scheme, all the models are treated equally and no significant improvement is observed in the performance over the best performing individual model.
AB - This paper presents our contributions to the MediaEval 2021 task namely”WaterMM: Water Quality in Social Multimedia”. The task aims at analyzing social media posts relevant to water quality with particular focus on the aspects like watercolor, smell, taste, and related illnesses. To this aim, a multimodal dataset containing both textual and visual information along with meta-data is provided. Considering the quality and quantity of available content, we mainly focus on textual information by employing three different models individually and jointly in a late-fusion manner. These models include (i) Bidirectional Encoder Representations from Transformers (BERT), (ii) Robustly Optimized BERT Pre-training Approach (XLM-RoBERTa), and a (iii) custom Long short-term memory (LSTM) model obtaining an overall F1-score of 0.794, 0.717, 0.663 on the official test set, respectively. In the fusion scheme, all the models are treated equally and no significant improvement is observed in the performance over the best performing individual model.
UR - http://www.scopus.com/inward/record.url?scp=85137041779&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85137041779
SN - 1613-0073
VL - 3181
JO - CEUR Workshop Proceedings
JF - CEUR Workshop Proceedings
T2 - MediaEval 2021 Workshop, MediaEval 2021
Y2 - 13 December 2021 through 15 December 2021
ER -