These are the resources and demos associated with the tutorial "Hate speech detection, mitigation and beyond" at ICWSM 2021 and AAAI 2022 are noted here.
Social media sites such as Twitter and Facebook have connected billions of people and given the opportunity to the users to share their ideas and opinions instantly. That being said, there are several ill consequences as well such as online harassment, trolling, cyber-bullying, fake news, and hate speech. Out of these, hate speech presents a unique challenge as it is deep engraved into our society and is often linked with offline violence. Social media platforms rely on local moderators to identify hate speech and take necessary action, but with a prolific increase in such content over the social media many are turning toward automated hate speech detection and mitigation systems. This shift brings several challenges on the plate, and hence, is an important avenue to explore for the computation social science community.
- Our papers are accepted in top conferences like AAAI, WWW, CSCW, ICWSM, WebSci. Link to the papers here
- We have open sourced our codes and datasets under a single github organisation - hate-alert for the future research in this domain
- We have stored different transformers models in huggingface.co. Link to hatealert organisation
- Dataset from our recent accepted paper in AAAI - "Hatexplain:A Benchmark Dataset for Explainable Hate Speech Detection" is also stored in the huggingface datsets forum
- We also participate in several hate speech shared tasks, winning many of them - hatealert@DLTEACL, hateminers@AMI, hatemonitors@HASOC and coming under 1% in hatealert@Hatememe detection by Facebook AI.
- Notion page containing hate speech papers.
- A dataset resource created and maintained by Leon Derczynski and Bertie Vidgen. Click the link here
- This resource collates all the resources and links used in this information hub, for both teachers and young people. Click the link here
We also provide some demos for the social scientists so that our opensource models can be used. Please provide feedback in the issues.
- Multlingual abuse predictor - This presents a suite of models which try to predict abuse in different languages. Different models are built upon the dataset found from that language. You can upload a file in the specified format and get back the predicitions of these models.
- Rationale predictor demo - This is a model trained using rationale and classifier head. Along with predicting the abusive or non-abusive label, it can also predict the rationales i.e. parts of text which are abusive according to the model.
- Counter speech detection demo - These are some of the models which can detect counter speech. These models are simple in nature. Link to the original github repository
🚨 Check the individual colab demos to learn more about the how to use these tools. These models might carry potential biases, hence should be used with appropriate caution. 🚨