CONTENT MODERATION FRAMEWORK FOR THE LLM-BASED RECOMMENDATION SYSTEMS

Authors

  • Rohan Singh Rajput Machine Learning, Headspace, Los Angeles, California, United States of America. Author
  • Sarthik Shah College of Business Administration, University of Illinois, Chicago, Illinois, United States of America. Author
  • Shantanu Neema Department of Computer Science, Georgia Institute of Technology, Atlanta, Georgia, United States of America. Author

Keywords:

Recommender System, Content Moderation, Large Language Model, Information Retrieval, Machine Learning

Abstract

Recommendation systems are integral to modern applications, personalization is bound to enhanced by Large Language Models (LLMs). However, their integration to these systems poses challenges, particularly in content moderation, essential for maintaining a safe user environment. Content moderation ensures relevance, trustworthiness, and safety by filtering inappropriate content. In past, some examples of AI backed tools raised concern about potential biases and legal implications. This paper explores the intersection of LLMs and recommendation systems, emphasizing the critical role of content moderation in mitigating risks. This work delves into historical content moderation papers, highlighting challenges in balancing contextual understanding, diversity, and compliance. This paper emphasizes the need for a tailored approach to address the dynamic nature of LLM-generated content for specific organization needs. This is an attempt to introduce a novel framework for LLM-generated content moderation, adapting to diverse business scopes, considering unique boundaries and dynamic nature of recommendations. This paper contributes a comprehensive exploration of intersection between LLMs, recommendation systems, and content moderation. The proposed framework aims to fill existing gaps, offering a dynamic solution for the intricate challenges posed by LLM-generated content in the rapidly evolving landscape of AI applications.

References

Hong S., Hyoung Kim S. H. (2016). Political polarization on twitter: Implications for the use of social media in digital governments, Vol 33, Issue 4, pp 777-782, ISSN 0740-624X, https://doi.org/10.1016/j.giq.2016.04.007

Wang W., Huang J., Chen C., Gu J. He P., Lyu M. R. (2023). An Image is Worth a Thousand Toxic Words: A Metamorphic Testing Framework for Content Moderation Software, 38th IEEE/ACM International Conference on Automated Software Engineering (ASE)

Wang W., Lin X., Feng F., He X., Chual T. (2023). Generative Recommendation: Towards Next-generation Recommender Paradigm, National University of Singapore, University of Science and Technology of China

Memarian, B., & Doleck, T. (2023). Fairness, accountability, transparency, and ethics (FATE) in artificial intelligence (AI) and higher education: A systematic review. Computers and Education: Artificial Intelligence, 5, Article 100152. https://doi.org/10.1016/j.caeai.2023.100152

Del Sesto R.W., Kwon T. (2023). The United States Approach to AI Regulartion: Key Considerations for Companies, Morgan Lewis Law firm

Buccino J. (2023). Guarding the AI frontier: A proposal for federal regulation, Federal Times Special Multimedia Report

Henin, C., Le Métayer, D. (2021). Towards a Framework for Challenging ML-Based Decisions. In: Sarkadi, S., Wright, B., Masters, P., McBurney, P. (eds) Deceptive AI. DeceptECAI DeceptAI 2020 2021. Communications in Computer and Information Science, vol 1296. Springer, Cham. https://doi.org/10.1007/978-3-030-91779-1_10

Gorwa, R., Binns, R., & Katzenbach, C. (2020). Algorithmic content moderation: Technical and political challenges in the automation of platform governance. Big Data & Society, 7(1). https://doi.org/10.1177/2053951719897945

Geiger, Christophe and Jütte, Bernd Justin, Towards a Virtuous Legal Framework for Content Moderation by Digital Platforms in the EU? The Commission’s Guidance on Article 17 CDSM Directive in the light of the YouTube/Cyando judgement and the AG’s Opinion in C-401/19 (July 18, 2021). European Intellectual Property Review (2021), Vol. 43, Issue 10, pp. 625-635., http://dx.doi.org/10.2139/ssrn.3889049

Wang, W., Huang, J. T., Wu, W., Zhang, J., Huang, Y., Li, S., Lyu, M. R. (2023). MTTM: Metamorphic testing for textual content moderation software. 45th International Conference on Software Engineering (ICSE) (pp. 2387-2399). IEEE.

Iriondo R. (2018). Amazon Scraps Secret AI Recruiting Engine that Showed Bias Against Women, Carnegie Mellon University’s News Archive

Lee D. (2016). Tay: Microsoft issues apology over racist chatbot fiasco, BBC, https://www.bbc.com/news/technology-35902104

Walsh D. (2022). Study: Social Media use Linked to Decline in Mental Health, Slone School of Management, Massachusetts Institute of Technology

Larson J., Mattu S., Kirchner L., Angwin J. (2016). How we Analyzed the COMPAS Recidivism Algorithm, ProPublica Investigative Journalism in the Public Interest

Downloads

Published

2023-12-02