My research project
Why do you hate me? Interpreting and explaining automatic classifications of hate in social media
There are numerous examples where hateful speech has been found online and several engines have been developed, focusing for example on racism, misogyny, etc. While they claim to work well on well defined datasets, it is not often clear whether a text is hateful or 'merely offensive'. This project aims to develop new methods to take the context in which a possible hateful utterance is found on social media, and provide interpretations or explanations helping to understand whether something is hateful.