Context-Aware Models for Text Classification in Sensitive Content Detection

Authors

  • Shashishekhar Ramagundam  

DOI:

https://doi.org/10.32628/IJSRSET22970

Keywords:

AUC, BERT, Deep Learning, Embeddings, Hate Speech.

Abstract

Sensitive content detection plays a pivotal role in ensuring the safety and integrity of digital platforms, especially with the increasing volume of user-generated content. Traditional models for content moderation often rely on keyword-based filtering systems that detect explicit offensive terms but fail to identify more subtle forms of harmful content where context plays a significant role. This paper presents a context-aware model for detecting sensitive content that integrates contextual embeddings from transformer-based models like BERT, coupled with deep learning techniques. Our proposed model leverages the power of contextual information, allowing it to understand the nuanced meaning behind text based on its surrounding words and context. The model was evaluated using the Hate Speech Dataset, and our results show a significant improvement in the detection of sensitive content compared to traditional rule-based and keyword-based models. Specifically, the context-aware model achieved a maximum accuracy of 88%, while the baseline rule-based model reached only 70% accuracy. By focusing on context, our approach improves the accuracy, recall, and precision in identifying not only direct hate speech but also more subtle forms of cyberbullying, harassment, and inappropriate language. The proposed method demonstrates the potential of context-aware models in enhancing content moderation, ensuring safer online interactions and contributing to more robust, scalable solutions for sensitive content detection.

References

  1. Waseem, Z., et al., “Hateful symbols or hateful people? Predicting hate speech on the World Wide Web,” Proceedings of the First Workshop on NLP and Computational Social Science, 2017.
  2. Davidson, T., et al., “Automated hate speech detection and the problem of offensive language,” Proceedings of the International Conference on ICWSM, 2017.
  3. Zhang, L., et al., “Detecting hate speech in social media,” Proceedings of the International Conference on Data Mining, 2017.
  4. Zhao, Q., et al., “Deep learning for content moderation in social media,” Proceedings of the Annual Meeting of the Association for Computational Linguistics, 2018.
  5. Kim, Y., et al., “Using machine learning for hate speech detection,” Proceedings of the Workshop on NLP and Social Media, 2017.
  6. Salama, M., et al., “Neural networks for detecting offensive language in social media,” Proceedings of the International Conference on Artificial Intelligence, 2018.
  7. Chen, J., et al., “Exploring deep learning for content moderation,” Proceedings of the Web Conference, 2017.
  8. Jha, S., et al., “Applying machine learning models to detect indirect hate speech,” Proceedings of the International Conference on NLP, 2018.
  9. Peters, M. E., et al., “Deep contextualized word representations,” Proceedings of the NAACL-HLT, 2018.
  10. Devlin, J., et al., “BERT: Pre-training of deep bidirectional transformers for language understanding,” Proceedings of NAACL, 2018.
  11. Gao, L., et al., “Context-aware text classification using transformer models,” Proceedings of the International Conference on NLP, 2018.
  12. Lan, Y., et al., “Contextual understanding for social media content moderation,” Proceedings of ACL, 2018.
  13. Lan, Y., et al., “Multitask learning for content moderation in social media,” Proceedings of the International Conference on Machine Learning, 2017.
  14. Kumar, M., et al., “Using multitask learning for improved hate speech detection,” Proceedings of the International Conference on AI, 2018.
  15. Li, J., et al., “Combining deep learning and rule-based methods for content moderation,” Proceedings of ICMLA, 2018.
  16. Barro, C., et al., “A survey of neural network models in content moderation,” Journal of AI and Big Data, vol. 8, no. 1, 2018.
  17. Tan, Q., et al., “Deep learning for nuanced hate speech detection,” Proceedings of the International Conference on NLP and Deep Learning, 2018.
  18. Binns, R., et al., “Optimizing NLP models for detecting sensitive content in real-time,” Proceedings of the Web Conference, 2017.
  19. Zhang, H., et al., “Combining semantic understanding and rule-based filters for content moderation,” Proceedings of the ICWSM, 2018.
  20. Waseem, Z., Thorne, J. and Bingel, J., “Bridging the gaps: Multi task learning for domain transfer of hate speech detection”, Online harassment, pp.29-55, 2018.

Downloads

Published

2019-02-22

Issue

Section

Research Articles

How to Cite

[1]
Shashishekhar Ramagundam "Context-Aware Models for Text Classification in Sensitive Content Detection" International Journal of Scientific Research in Science, Engineering and Technology (IJSRSET), Print ISSN : 2395-1990, Online ISSN : 2394-4099, Volume 6, Issue 1, pp.630-639, January-February-2019. Available at doi : https://doi.org/10.32628/IJSRSET22970