Analyzing Norm Violations in Live-Stream Chat. Moon, J., Lee, D., Cho, H., Jin, W., Park, C., Kim, M., May, J., Pujara, J., & Park, S. In Bouamor, H., Pino, J., & Bali, K., editors, Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 852–868, Singapore, December, 2023. Association for Computational Linguistics. Paper doi abstract bibtex 3 downloads Toxic language, such as hate speech, can deter users from participating in online communities and enjoying popular platforms. Previous approaches to detecting toxic language and norm violations have been primarily concerned with conversations from online forums and social media, such as Reddit and Twitter. These approaches are less effective when applied to conversations on live-streaming platforms, such as Twitch and YouTube Live, as each comment is only visible for a limited time and lacks a thread structure that establishes its relationship with other comments. In this work, we share the first NLP study dedicated to detecting norm violations in conversations on live-streaming platforms. We define norm violation categories in live-stream chats and annotate 4,583 moderated comments from Twitch. We articulate several facets of live-stream data that differ from other forums, and demonstrate that existing models perform poorly in this setting. By conducting a user study, we identify the informational context humans use in live-stream moderation, and train models leveraging context to identify norm violations. Our results show that appropriate contextual information can boost moderation performance by 35%.
@inproceedings{moon-etal-2023-analyzing,
title = "Analyzing Norm Violations in Live-Stream Chat",
author = "Moon, Jihyung and
Lee, Dong-Ho and
Cho, Hyundong and
Jin, Woojeong and
Park, Chan and
Kim, Minwoo and
May, Jonathan and
Pujara, Jay and
Park, Sungjoon",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.emnlp-main.55",
doi = "10.18653/v1/2023.emnlp-main.55",
pages = "852--868",
abstract = "Toxic language, such as hate speech, can deter users from participating in online communities and enjoying popular platforms. Previous approaches to detecting toxic language and norm violations have been primarily concerned with conversations from online forums and social media, such as Reddit and Twitter. These approaches are less effective when applied to conversations on live-streaming platforms, such as Twitch and YouTube Live, as each comment is only visible for a limited time and lacks a thread structure that establishes its relationship with other comments. In this work, we share the first NLP study dedicated to detecting norm violations in conversations on live-streaming platforms. We define norm violation categories in live-stream chats and annotate 4,583 moderated comments from Twitch. We articulate several facets of live-stream data that differ from other forums, and demonstrate that existing models perform poorly in this setting. By conducting a user study, we identify the informational context humans use in live-stream moderation, and train models leveraging context to identify norm violations. Our results show that appropriate contextual information can boost moderation performance by 35{\%}.",
}
Downloads: 3
{"_id":"itqu6oCyAzsAr5Am3","bibbaseid":"moon-lee-cho-jin-park-kim-may-pujara-etal-analyzingnormviolationsinlivestreamchat-2023","author_short":["Moon, J.","Lee, D.","Cho, H.","Jin, W.","Park, C.","Kim, M.","May, J.","Pujara, J.","Park, S."],"bibdata":{"bibtype":"inproceedings","type":"inproceedings","title":"Analyzing Norm Violations in Live-Stream Chat","author":[{"propositions":[],"lastnames":["Moon"],"firstnames":["Jihyung"],"suffixes":[]},{"propositions":[],"lastnames":["Lee"],"firstnames":["Dong-Ho"],"suffixes":[]},{"propositions":[],"lastnames":["Cho"],"firstnames":["Hyundong"],"suffixes":[]},{"propositions":[],"lastnames":["Jin"],"firstnames":["Woojeong"],"suffixes":[]},{"propositions":[],"lastnames":["Park"],"firstnames":["Chan"],"suffixes":[]},{"propositions":[],"lastnames":["Kim"],"firstnames":["Minwoo"],"suffixes":[]},{"propositions":[],"lastnames":["May"],"firstnames":["Jonathan"],"suffixes":[]},{"propositions":[],"lastnames":["Pujara"],"firstnames":["Jay"],"suffixes":[]},{"propositions":[],"lastnames":["Park"],"firstnames":["Sungjoon"],"suffixes":[]}],"editor":[{"propositions":[],"lastnames":["Bouamor"],"firstnames":["Houda"],"suffixes":[]},{"propositions":[],"lastnames":["Pino"],"firstnames":["Juan"],"suffixes":[]},{"propositions":[],"lastnames":["Bali"],"firstnames":["Kalika"],"suffixes":[]}],"booktitle":"Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing","month":"December","year":"2023","address":"Singapore","publisher":"Association for Computational Linguistics","url":"https://aclanthology.org/2023.emnlp-main.55","doi":"10.18653/v1/2023.emnlp-main.55","pages":"852–868","abstract":"Toxic language, such as hate speech, can deter users from participating in online communities and enjoying popular platforms. Previous approaches to detecting toxic language and norm violations have been primarily concerned with conversations from online forums and social media, such as Reddit and Twitter. These approaches are less effective when applied to conversations on live-streaming platforms, such as Twitch and YouTube Live, as each comment is only visible for a limited time and lacks a thread structure that establishes its relationship with other comments. In this work, we share the first NLP study dedicated to detecting norm violations in conversations on live-streaming platforms. We define norm violation categories in live-stream chats and annotate 4,583 moderated comments from Twitch. We articulate several facets of live-stream data that differ from other forums, and demonstrate that existing models perform poorly in this setting. By conducting a user study, we identify the informational context humans use in live-stream moderation, and train models leveraging context to identify norm violations. Our results show that appropriate contextual information can boost moderation performance by 35%.","bibtex":"@inproceedings{moon-etal-2023-analyzing,\n title = \"Analyzing Norm Violations in Live-Stream Chat\",\n author = \"Moon, Jihyung and\n Lee, Dong-Ho and\n Cho, Hyundong and\n Jin, Woojeong and\n Park, Chan and\n Kim, Minwoo and\n May, Jonathan and\n Pujara, Jay and\n Park, Sungjoon\",\n editor = \"Bouamor, Houda and\n Pino, Juan and\n Bali, Kalika\",\n booktitle = \"Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing\",\n month = dec,\n year = \"2023\",\n address = \"Singapore\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2023.emnlp-main.55\",\n doi = \"10.18653/v1/2023.emnlp-main.55\",\n pages = \"852--868\",\n abstract = \"Toxic language, such as hate speech, can deter users from participating in online communities and enjoying popular platforms. Previous approaches to detecting toxic language and norm violations have been primarily concerned with conversations from online forums and social media, such as Reddit and Twitter. These approaches are less effective when applied to conversations on live-streaming platforms, such as Twitch and YouTube Live, as each comment is only visible for a limited time and lacks a thread structure that establishes its relationship with other comments. In this work, we share the first NLP study dedicated to detecting norm violations in conversations on live-streaming platforms. We define norm violation categories in live-stream chats and annotate 4,583 moderated comments from Twitch. We articulate several facets of live-stream data that differ from other forums, and demonstrate that existing models perform poorly in this setting. By conducting a user study, we identify the informational context humans use in live-stream moderation, and train models leveraging context to identify norm violations. Our results show that appropriate contextual information can boost moderation performance by 35{\\%}.\",\n}\n\n","author_short":["Moon, J.","Lee, D.","Cho, H.","Jin, W.","Park, C.","Kim, M.","May, J.","Pujara, J.","Park, S."],"editor_short":["Bouamor, H.","Pino, J.","Bali, K."],"key":"moon-etal-2023-analyzing","id":"moon-etal-2023-analyzing","bibbaseid":"moon-lee-cho-jin-park-kim-may-pujara-etal-analyzingnormviolationsinlivestreamchat-2023","role":"author","urls":{"Paper":"https://aclanthology.org/2023.emnlp-main.55"},"metadata":{"authorlinks":{}},"downloads":3},"bibtype":"inproceedings","biburl":"https://jonmay.github.io/webpage/cutelabname/cutelabname.bib","dataSources":["j3Qzx9HAAC6WtJDHS","5eM3sAccSEpjSDHHQ","8coo8cZMjGAhtHyqT"],"keywords":[],"search_terms":["analyzing","norm","violations","live","stream","chat","moon","lee","cho","jin","park","kim","may","pujara","park"],"title":"Analyzing Norm Violations in Live-Stream Chat","year":2023,"downloads":5}