CONTENT MODERATION FRAMEWORK FOR THE LLM-BASED RECOMMENDATION SYSTEMS
Keywords:
Recommender System, Content Moderation, Large Language Model, Information Retrieval, Machine LearningAbstract
Recommendation systems are integral to modern applications, personalization is bound to enhanced by Large Language Models (LLMs). However, their integration to these systems poses challenges, particularly in content moderation, essential for maintaining a safe user environment. Content moderation ensures relevance, trustworthiness, and safety by filtering inappropriate content. In past, some examples of AI backed tools raised concern about potential biases and legal implications. This paper explores the intersection of LLMs and recommendation systems, emphasizing the critical role of content moderation in mitigating risks. This work delves into historical content moderation papers, highlighting challenges in balancing contextual understanding, diversity, and compliance. This paper emphasizes the need for a tailored approach to address the dynamic nature of LLM-generated content for specific organization needs. This is an attempt to introduce a novel framework for LLM-generated content moderation, adapting to diverse business scopes, considering unique boundaries and dynamic nature of recommendations. This paper contributes a comprehensive exploration of intersection between LLMs, recommendation systems, and content moderation. The proposed framework aims to fill existing gaps, offering a dynamic solution for the intricate challenges posed by LLM-generated content in the rapidly evolving landscape of AI applications.
References
Hong S., Hyoung Kim S. H. (2016). Political polarization on twitter: Implications for the use of social media in digital governments, Vol 33, Issue 4, pp 777-782, ISSN 0740-624X, https://doi.org/10.1016/j.giq.2016.04.007
Wang W., Huang J., Chen C., Gu J. He P., Lyu M. R. (2023). An Image is Worth a Thousand Toxic Words: A Metamorphic Testing Framework for Content Moderation Software, 38th IEEE/ACM International Conference on Automated Software Engineering (ASE)
Wang W., Lin X., Feng F., He X., Chual T. (2023). Generative Recommendation: Towards Next-generation Recommender Paradigm, National University of Singapore, University of Science and Technology of China
Memarian, B., & Doleck, T. (2023). Fairness, accountability, transparency, and ethics (FATE) in artificial intelligence (AI) and higher education: A systematic review. Computers and Education: Artificial Intelligence, 5, Article 100152. https://doi.org/10.1016/j.caeai.2023.100152
Del Sesto R.W., Kwon T. (2023). The United States Approach to AI Regulartion: Key Considerations for Companies, Morgan Lewis Law firm
Buccino J. (2023). Guarding the AI frontier: A proposal for federal regulation, Federal Times Special Multimedia Report
Henin, C., Le Métayer, D. (2021). Towards a Framework for Challenging ML-Based Decisions. In: Sarkadi, S., Wright, B., Masters, P., McBurney, P. (eds) Deceptive AI. DeceptECAI DeceptAI 2020 2021. Communications in Computer and Information Science, vol 1296. Springer, Cham. https://doi.org/10.1007/978-3-030-91779-1_10
Gorwa, R., Binns, R., & Katzenbach, C. (2020). Algorithmic content moderation: Technical and political challenges in the automation of platform governance. Big Data & Society, 7(1). https://doi.org/10.1177/2053951719897945
Geiger, Christophe and Jütte, Bernd Justin, Towards a Virtuous Legal Framework for Content Moderation by Digital Platforms in the EU? The Commission’s Guidance on Article 17 CDSM Directive in the light of the YouTube/Cyando judgement and the AG’s Opinion in C-401/19 (July 18, 2021). European Intellectual Property Review (2021), Vol. 43, Issue 10, pp. 625-635., http://dx.doi.org/10.2139/ssrn.3889049
Wang, W., Huang, J. T., Wu, W., Zhang, J., Huang, Y., Li, S., Lyu, M. R. (2023). MTTM: Metamorphic testing for textual content moderation software. 45th International Conference on Software Engineering (ICSE) (pp. 2387-2399). IEEE.
Iriondo R. (2018). Amazon Scraps Secret AI Recruiting Engine that Showed Bias Against Women, Carnegie Mellon University’s News Archive
Lee D. (2016). Tay: Microsoft issues apology over racist chatbot fiasco, BBC, https://www.bbc.com/news/technology-35902104
Walsh D. (2022). Study: Social Media use Linked to Decline in Mental Health, Slone School of Management, Massachusetts Institute of Technology
Larson J., Mattu S., Kirchner L., Angwin J. (2016). How we Analyzed the COMPAS Recidivism Algorithm, ProPublica Investigative Journalism in the Public Interest