Bilingual Cyber-aggression Detection on Social Media using LSTM Autoencoder
Publication date
2021-07Keyword
Long short term memoryOnline social networks
Cyberbullying
Social media
Cyber-aggression
Detection
Rights
(c) 2021 SpringerNature. Full-text reproduced in accordance with the publisher's self-archiving policy.Peer-Reviewed
YesOpen Access status
openAccess
Metadata
Show full item recordAbstract
Cyber-aggression is an offensive behaviour attacking people based on race, ethnicity, religion, gender, sexual orientation, and other traits. It has become a major issue plaguing the online social media. In this research, we have developed a deep learning-based model to identify different levels of aggression (direct, indirect and no aggression) in a social media post in a bilingual scenario. The model is an autoencoder built using the LSTM network and trained with non-aggressive comments only. Any aggressive comment (direct or indirect) will be regarded as an anomaly to the system and will be marked as Overtly (direct) or Covertly (indirect) aggressive comment depending on the reconstruction loss by the autoencoder. The validation results on the dataset from two popular social media sites: Facebook and Twitter with bilingual (English and Hindi) data outperformed the current state-of-the-art models with improvements of more than 11% on the test sets of the English dataset and more than 6% on the test sets of the Hindi dataset.Version
Accepted manuscriptCitation
Kumari K, Singh JP, Dwivedi YK et al (2021) Bilingual Cyber-aggression Detection on Social Media using LSTM Autoencoder. Soft Computing. 25: 8999-9012.Link to Version of Record
https://doi.org/10.1007/s00500-021-05817-yType
Articleae974a485f413a2113503eed53cd6c53
https://doi.org/10.1007/s00500-021-05817-y