Multilingual detection of hate speech against immigrants and women in Twitter (hatEval)
Hate Speech is commonly defined as any communication that disparages a person or a group on the basis of some characteristic such as race, color, ethnicity, gender, sexual orientation, nationality, religion, or other characteristics. Given the huge amount of user-generated contents on the Web, and in particular on social media, the problem of detecting, and therefore possibly limit the Hate Speech diffusion, is becoming fundamental, for instance for fighting against misogyny and xenophobia.
The proposed task consists in Hate Speech detection in Twitter but featured by two specific different targets, immigrants and women, in a multilingual perspective, for Spanish and English.
The task will be articulated around two related subtasks for each of the involved languages: a basic task about Hate Speech, and another one where fine-grained features of hateful contents will be investigated in order to understand how existing approaches may deal with the identification of especially dangerous forms of hate, i.e. those where the incitement is against an individual rather than against a group of people, and where an aggressive behavior of the author can be identified as a prominent feature of the expression of hate. Participants will be asked to identify, on the one hand, if the target of hate is a single human or a group of persons, on the other hand, if the message author intends to be aggressive, harmful, or even to incite, in various forms, to violent acts against the target.
Important dates
Join the hatEval mailing group: semeval2019-task5-hateval[at]googlegroups.com
Organizers:
Evaluation
For the evaluation of the results of task A and B different strategies and metrics are applied in order to allow for more fine-grained scores.
TASK A.
Systems will be evaluated using standard evaluation metrics, including accuracy, precision, recall and F1-score. The submissions will be ranked by F1-score.
The metrics will be computed as follows:
Systems will be evaluated on the basis of two criteria: partial match and exact match.
where Zi denotes the set of labels predicted for the i-th instance and I is the indicator function.
The submissions will be ranked by EMR. This choice is motivated by the willingness of capturing the most difficult task of capturing the entire phenomena, and therefore to identify the most dangerous behaviours against the targets.
The evaluation script is available in this GitHub repository:
https://github.com/msang/hateval/tree/master/SemEval2019-Task5/evaluation
During the Practice phase, the prediction files submitted by participants to the task page will be evaluated for the task A, and for demonstration purposes only; if participants wish to test the script on prediction files for task B as well, they could use the version available in the GitHub repository.
For the Development and Evaluation phases, the script will provide a complete evaluation for each language and task for any submitted file, provided that the latter meet the submission requirements (see Submission Instructions).
The official hatEval evaluation script takes one single prediction file as input, for each task and for each language, that MUST be a TSV file structured as follows:
id[tab]{0|1}
e.g.
101[tab]1
102[tab]0
103[tab]1
id[tab]{0|1}[tab]{0|1}[tab]{0|1}
e.g.
101[tab]1[tab]1[tab]1
102[tab]0[tab]0[tab]0
103[tab]1[tab]1[tab]0
104[tab]1[tab]0[tab]0
105[tab]1[tab]0[tab]1
Contrary to the trial and training set, the submission files do NOT have the header in the first line.
When submitting predictions to the task page in Codalab, one single file should be uploaded for each task and language, as a zip-compressed file, and it should be named according to the language and task predictions are submitted for, therefore:
NOTE:
For the Practice phase, more than one submission is allowed, BUT for the task A only. While during the Development and Evaluation phases, participants are free to submit their system's predictions for each language and task separately.
For the Development phase participants will be able to make more than one submission for each language and task, while for the Evaluation phase, a maximum of 2 submissions has been set for each language and for both task A and B, but please note that only the final valid one is taken as the official submission for the competition.
Terms and conditions
By submitting results to this competition, you consent to the public release of your scores at the SemEval-2019 workshop and in the associated proceedings, at the task organizers' discretion. Scores may include, but are not limited to, automatic and manual quantitative judgements, qualitative judgements, and such other metrics as the task organizers see fit. You accept that the ultimate decision of metric choice and score value is that of the task organizers.
You further agree that the task organizers are under no obligation to release scores and that scores may be withheld if it is the task organizers' judgement that the submission was incomplete, erroneous, deceptive, or violated the letter or spirit of the competition's rules. Inclusion of a submission's scores is not an endorsement of a team or individual's submission, system, or science.
You further agree that your system may be named according to the team name provided at the time of submission, or to a suitable shorthand as determined by the task organizers.
You agree not to redistribute the test data except in the manner prescribed by its licence.
Start: Aug. 20, 2018, midnight
Description: Trial data available.
Start: Sept. 17, 2018, midnight
Description: English dataset for task A available for training. More than one submission allowed in this phase.
Start: Sept. 17, 2018, midnight
Description: Spanish dataset for task A available for training. More than one submission allowed in this phase.
Start: Sept. 17, 2018, midnight
Description: Spanish dataset for task B available for training. More than one submission allowed in this phase.
Start: Sept. 17, 2018, midnight
Description: English dataset for task B available for training. More than one submission allowed in this phase.
Start: Jan. 10, 2019, midnight
Description: English test set available for task A. Up to 2 submissions are allowed, but only the final valid one is taken as the official submission for the competition.
Start: Jan. 10, 2019, midnight
Description: Spanish test set available for task A. Up to 2 submissions are allowed, but only the final valid one is taken as the official submission for the competition.
Start: Jan. 10, 2019, midnight
Description: English test set available for task B. Up to 2 submissions are allowed, but only the final valid one is taken as the official submission for the competition.
Start: Jan. 10, 2019, midnight
Description: Spanish test set available for task B. Up to 2 submissions are allowed, but only the final valid one is taken as the official submission for the competition.
Never
You must be logged in to participate in competitions.
Sign In