Official repository of AMI, the shared task on Automatic Misogyny Identification at Evalita 2020.
Training and testing set available filling in this form; an email notification will be sent with instructions and details about how to download the data.
For more information on the task, see the guidelines available in this repository, and the task web page.
To install the required python packages, run:
pip install -r requirements.txt
In the case of problems, try to run pip install --upgrade pip setuptools
first.
The evaluation script can be used for evaluating the results both of Subtask A and B given a zip file containing submission files as input. For submission files formats check the AMI 2020 Guidelines.
For running the evaluation script for Subtask A, you can run something like the following:
python evaluation_submission.py \
--submission_path teamName.zip \
--gold_path_raw AMI2020_TrainingSet/AMI2020_TrainingSet.tsv \
--task a \
--output_path result.tsv
For running the evaluation script for Subtask B, you can run something like the following:
python evaluation_submission.py \
--submission_path teamName.zip \
--gold_path_raw AMI2020_TrainingSet/AMI2020_TrainingSet.tsv \
--gold_path_synt AMI2020_TrainingSet/AMI2020_training_synt.tsv \
--identityterms_path AMI2020_TrainingSet/AMI2020_training_identityterms.txt \
--task b \
--output_path result.tsv
If you use this in a research work please cite this paper:
@inproceedings{Fersini2020,
author = {Elisabetta Fersini, Debora Nozza, Paolo Rosso},
title = {AMI @ EVALITA2020: Automatic Misogyny Identification},
booktitle = {{Proceedings of the 7th evaluation campaign of Natural Language Processing and Speech tools for Italian (EVALITA 2020)}},
editor = {Basile, Valerio and Croce, Danilo and Di Maro, Maria and Passaro, Lucia C.},
year = {2020},
publisher = {CEUR.org},
address = {Online}
}
If you find issues on the evaluation script, please contact Debora Nozza: Twitter | Github | Webpage