Analyzing Norm Violations in Live-Stream Chat. Moon, J., Lee, D., Cho, H., Jin, W., Park, C., Kim, M., May, J., Pujara, J., & Park, S. In Bouamor, H., Pino, J., & Bali, K., editors, Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 852–868, Singapore, December, 2023. Association for Computational Linguistics.
Analyzing Norm Violations in Live-Stream Chat [link]Paper  doi  abstract   bibtex   3 downloads  
Toxic language, such as hate speech, can deter users from participating in online communities and enjoying popular platforms. Previous approaches to detecting toxic language and norm violations have been primarily concerned with conversations from online forums and social media, such as Reddit and Twitter. These approaches are less effective when applied to conversations on live-streaming platforms, such as Twitch and YouTube Live, as each comment is only visible for a limited time and lacks a thread structure that establishes its relationship with other comments. In this work, we share the first NLP study dedicated to detecting norm violations in conversations on live-streaming platforms. We define norm violation categories in live-stream chats and annotate 4,583 moderated comments from Twitch. We articulate several facets of live-stream data that differ from other forums, and demonstrate that existing models perform poorly in this setting. By conducting a user study, we identify the informational context humans use in live-stream moderation, and train models leveraging context to identify norm violations. Our results show that appropriate contextual information can boost moderation performance by 35%.
@inproceedings{moon-etal-2023-analyzing,
    title = "Analyzing Norm Violations in Live-Stream Chat",
    author = "Moon, Jihyung  and
      Lee, Dong-Ho  and
      Cho, Hyundong  and
      Jin, Woojeong  and
      Park, Chan  and
      Kim, Minwoo  and
      May, Jonathan  and
      Pujara, Jay  and
      Park, Sungjoon",
    editor = "Bouamor, Houda  and
      Pino, Juan  and
      Bali, Kalika",
    booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
    month = dec,
    year = "2023",
    address = "Singapore",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2023.emnlp-main.55",
    doi = "10.18653/v1/2023.emnlp-main.55",
    pages = "852--868",
    abstract = "Toxic language, such as hate speech, can deter users from participating in online communities and enjoying popular platforms. Previous approaches to detecting toxic language and norm violations have been primarily concerned with conversations from online forums and social media, such as Reddit and Twitter. These approaches are less effective when applied to conversations on live-streaming platforms, such as Twitch and YouTube Live, as each comment is only visible for a limited time and lacks a thread structure that establishes its relationship with other comments. In this work, we share the first NLP study dedicated to detecting norm violations in conversations on live-streaming platforms. We define norm violation categories in live-stream chats and annotate 4,583 moderated comments from Twitch. We articulate several facets of live-stream data that differ from other forums, and demonstrate that existing models perform poorly in this setting. By conducting a user study, we identify the informational context humans use in live-stream moderation, and train models leveraging context to identify norm violations. Our results show that appropriate contextual information can boost moderation performance by 35{\%}.",
}

Downloads: 3