Authors :
Dhruvi Dineshbhai Patel; Vishnupant Potdar; Dr. Nagnath Biradar
Volume/Issue :
Volume 10 - 2025, Issue 7 - July
Google Scholar :
https://tinyurl.com/ypd9hc4w
Scribd :
https://tinyurl.com/mww22f97
DOI :
https://doi.org/10.38124/ijisrt/25jul233
Note : A published paper may take 4-5 working days from the publication date to appear in PlumX Metrics, Semantic Scholar, and ResearchGate.
Note : Google Scholar may take 30 to 40 days to display the article.
Abstract :
With the growing use of digital platforms, online bullying has become a serious and widespread issue that
significantly affects users’ mental and emotional well-being. Public figures such as influencers and celebrities face even
greater vulnerability due to their high visibility and constant exposure on online networks, especially after the sharp rise in
social media usage following the pandemic. To tackle this challenge, our work adopts a well-rounded strategy that examines
both the text and the expressive cues conveyed by emojis to detect cyberbullying. We utilize an array of machine learning
and deep learning models—namely, Support Vector Classifier, Logistic Regression, Random Forest, XGBoost, LSTM, Bi-
LSTM, GRU, and Bi-GRU to classify comments as bullying or non-bullying. Furthermore, we introduce a severity-based
scoring system that rates offensive text on a scale of 1 to 5. When a message crosses a predefined severity threshold—
determined by the safety standards of each platform—an automated recommendation to block the user is triggered. This
approach not only enables precise identification of harmful content but also provides a proactive mechanism to promote
safer online interactions.
Keywords :
Online Bullying Detection, Swear Words Dataset, TF-IDF With Random Forest, Severity Score.
References :
- J. Doe, R. Johnson, and S. Williams, "Understanding online bullying: A holistic Review," Journal of Cybersecurity, vol. 10, no. 3, pp. 123-145, 2023.
- Smith, B. Wilson, and C. Brown, "Online Harassment of Public Figures: A Case Study on Celebrity online bullying," in International Conference on Cybersecurity, pp. 56-67, 2023.
- K. Brown, J. Smith, L. Johnson, and M. Garcia, "Social Media Trends Post-Pandemic," Journal of Digital Communication and Social Media, vol. 8, no. 2, pp. 201-215, 2023.
- D. W. Hosmer, Jr., S. Lemeshow, and R. X. Sturdivant, "Applied logistic regression," John Wiley & Sons, 2013.
- L. Breiman, "Random forests," ML, vol. 45, no. 1, pp. 5-32, 2001.
- V. Vapnik, "The nature of statistical learning theory," Springer Science & Business Media, 1995.
- T. Chen and C. Guestrin, "Xgboost: A scalable and portable parallelized tree boosting framework," in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 785-794, 2016.
- S. Hochreiter and J. Schmidhuber, "Long short-term memory," Neural computation, vol. 9, no. 8, pp. 1735-1780, 1997.
- Graves, M. Schuster, and J. Schmidhuber, "Bidirectional recurrent neural networks," IEEE transactions on speech and audio processing, vol. 13, no. 5, pp. 793-801, 2005.
- K. Cho, B. Merrienboer, S. Gulcehre, D. Bahdanau, F. Boureau, Y. Bengio, and C. Baroni, "Learning phrase representations using recurrent neural networks," arXiv preprint arXiv:1406.1054, 2014.
- J. Chung, S. Gulcehre, K. Cho, and Y. Bengio, "Gated recurrent units with applications to language modeling and sentiment analysis," arXiv preprint arXiv:1412.3900, 2015.
- R. Gomez, J. Gibert, L. Gomez, and D. Karatzas, "Exploring Hate Speech Detection in Multimodal Publications," Computational Linguistics Review, vol. 20, no. 3, pp. 112-129, 2019.
- K. Reynolds, A. Kontostathis, and L. Edwards, "Using ML to detect online bullying," Journal of Social Networking, vol. 5, no. 2, pp. 112-125, 2011.
- S. Mitchell et al., "Cyber-aggression, online bullying, and Cyber-grooming: A Survey and Research Challenges," J. Cybersecurity Res., vol. 5, no. 2, pp. 123-145, 2019.
- J. Salminen et al., "Developing an online hate classifier for multiple online social networks," Journal of Computational Social Science, vol. 3, no. 2, pp. 333-348, 2020.
- P. Mishra, H. Yannakoudakis, and E. Shutova, "Tackling Online Abuse: A Survey of Automated Abuse Detection Methods," J. Nat. Lang. Process. Comput. Linguist., vol. 17, no. 3, pp. 215-238, 2019.
- N. Mehendale, K. Shah, C. Phadtare, and K. Rajpara, "Cyber Bullying Detection for Hindi- English Language Using
- ML," SSRN, 2022. [Online]. Available: https://ssrn.com/abstract=4116143.
- S. Neelakandan, M. Sridevi, S. Chandrasekaran, K. Murugeswari, A. K. Singh Pundir, R. Sridevi, and T. Bheema Lingaiah, "DL Approaches for online bullying Detection and Classification on Social Media," Hindawi Computational Intelligence and Neuroscience, vol. 2022, article ID 2163458, pp. 1-13, Jun. 11, 2022. DOI: 10.1155/2022/2163458.
- B. A. Talpur and D. O’Sullivan, “online bullying severity detection: a ML approach,” PLoS One, vol. 15,no. 10, Article ID e0240924, 2020.
- N. Yuvaraj, V. Chang, B. Gobinathan et al., “Automatic detection of online bullying using multi- feature based artificial intelligence with deep decision tree classification,” Computers & Electrical Engineering, vol. 92, Article ID 107186, 2021.
- Z. L. Chia, M. Ptaszynski, F. Masui, G. Leliwa, and M. Wroczynski, “ML and feature engineeringbased study into sarcasm and irony classification with application to online bullying detection,” Information Processing & Management, vol. 58, no. 4, Article ID102600, 2021.
- C. Van Hee et al., "Automatic detection of online bullying in social media text," PLoS ONE, vol. 13, no. 10, e0203794, 2018.
- Perera and P. Fernando, "Accurate online bullying Detection and Prevention on Social Media," Procedia Computer Science, vol. 181, pp. 605-611, 2021.
- D. Yin, Z. Xue, and L. Hong, “Detection of Harassment on Web 2.0,” p. 8, 2019.
- S. O. Sood, E. F. Churchill, and J. Antin, “Automatic identification of personal insults on social news sites,” J. Am. Soc. Inf. Sci. Technol., vol. 63, no. 2, pp. 270–285, Feb. 2012, doi: 10.1002/ asi.21690.
- Squicciarini, S. Rajtmajer, Y. Liu, and C. Griffin, “Identification and characterization of online bullying dynamics in an online social network,” in Proceedings of the 2015 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining 2015 - ASONAM ’15, Paris, France, 2015, pp. 280–285, doi: 10.1145/2808797.2809398.
- V. S. Chavan and Shylaja S S, “ML approach for detection of cyber-aggressive comments by peers on social media network,” in 2015 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Kochi, India, Aug. 2015, pp. 2354–2358, doi: 10.1109/ICACCI.2015.7275970.
- Z. Waseem and D. Hovy, "Hateful symbols or hateful people? Predictive features for hate speech detection on Twitter," in Proceedings of the NAACL Student Research Workshop, pp. 88-93, 2016.
- T. Davidson, D. Warmsley, M. Macy, and I. Weber, "Automated Hate Speech Detection and the Problem of Offensive Language," in Proceedings of the 11th International AAAI Conference on Web and Social Media,
- pp. 512-515, 2017. [Online]. Available: https://github.com/t- davidson/hate-speech-and-offensive- language.
With the growing use of digital platforms, online bullying has become a serious and widespread issue that
significantly affects users’ mental and emotional well-being. Public figures such as influencers and celebrities face even
greater vulnerability due to their high visibility and constant exposure on online networks, especially after the sharp rise in
social media usage following the pandemic. To tackle this challenge, our work adopts a well-rounded strategy that examines
both the text and the expressive cues conveyed by emojis to detect cyberbullying. We utilize an array of machine learning
and deep learning models—namely, Support Vector Classifier, Logistic Regression, Random Forest, XGBoost, LSTM, Bi-
LSTM, GRU, and Bi-GRU to classify comments as bullying or non-bullying. Furthermore, we introduce a severity-based
scoring system that rates offensive text on a scale of 1 to 5. When a message crosses a predefined severity threshold—
determined by the safety standards of each platform—an automated recommendation to block the user is triggered. This
approach not only enables precise identification of harmful content but also provides a proactive mechanism to promote
safer online interactions.
Keywords :
Online Bullying Detection, Swear Words Dataset, TF-IDF With Random Forest, Severity Score.