JIGSAW @ AMI AND HASPEEDE2: FINE-TUNING A PRE-TRAINED COMMENT-DOMAIN BERT MODEL
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
Jigsaw @ AMI and HaSpeeDe2: Fine-Tuning a Pre-Trained Comment-Domain BERT Model Alyssa Lees and Jeffrey Sorensen and Ian Kivlichan Google Jigsaw New York, NY (alyssalees|sorenj|kivlichan)@google.com Abstract labeled by crowd workers. Note that Perspec- tiveAPI actually hosts a number of different mod- The Google Jigsaw team produced els that each score different attributes. The under- submissions for two of the EVALITA lying technology and performance of these models 2020 (Basile et al., 2020) shared tasks, has evolved over time. based in part on the technology that pow- While Jigsaw has hosted three separate Kaggle ers the publicly available PerspectiveAPI competitions relevant to this competition (Jigsaw, comment evaluation service. We present a 2018; Jigsaw, 2019; Jigsaw, 2020) we have not basic description of our submitted results traditionally participated in academic evaluations. and a review of the types of errors that our system made in these shared tasks. 3 Related Work The models we build are based on the popular 1 Introduction BERT architecture (Devlin et al., 2019) with dif- The HaSpeeDe2 shared task consists of Italian so- ferent pre-training and fine-tuning approaches. cial media posts that have been labeled for hate In part, our submissions explore the importance speech and stereotypes. As Jigsaw’s participation of pre-training (Gururangan et al., 2020) in the was limited to the A and B tasks, we will be lim- context of toxicity and the various competition at- iting our analysis to that portion. The full details tributes. A core question is to what extent these of the dataset are available in the task guidelines domains overlap. Jigsaw’s customized models (Bosco et al., 2020). (used for the second HaSpeeDe2 submission, and The AMI task includes both raw (natural Twit- both AMI submissions) are pretrained on a set of ter) and synthetic (template-generated) datasets. one billion user-generated comments: this imparts The raw data consists of Italian tweets manually statistical information to the model about com- labelled and balanced according to misogyny and ments and conversations online. This model is fur- aggressiveness labels, while the synthetic data is ther fine-tuned on various toxicity attributes (toxi- labelled only for misogyny and is intended to city, severe toxicity, profanity, insults, identity at- measure the presence of unintended bias (Elisa- tacks, and threats), but it is unclear how well these betta Fersini, 2020). should align with the competition attributes. The descriptions of these attributes and how they were 2 Background collected from crowd workers can be found in the data descriptions for the Jigsaw Unintended Bias Jigsaw, a team within Google, develops the Per- in Toxicity Classification (Jigsaw, 2019) website. spectiveAPI machine learning comment scoring A second question studied in prior work is to system, which is used by numerous social media what extent training generalizes across languages companies and publishers. Our system is based (Pires et al., 2019; Wu and Dredze, 2019; Pa- on distillation and uses a convolutional neural- mungkas et al., 2020). The majority of our train- network to score individual comments according ing data is English comment data from a variety to several attributes using supervised training data of sources, while this competition is based on Ital- ian Twitter data. Though multilingual transfer has Copyright ©2020 for this paper by its authors. Use per- mitted under Creative Commons License Attribution 4.0 In- been studied in general contexts, less is known ternational (CC BY 4.0). about the specific cases of toxicity, hate speech,
misogyny, and harassment. This was one of the fo- gual teacher model (that is too large to practi- cuses of Jigsaw’s recent Kaggle competition (Jig- cally serve in production) to a smaller CNN. Using saw, 2020); i.e., what forms of toxicity are shared this large teacher model, we initially compared the across languages (and hence can be learned by EVALITA hate speech and stereotype annotations multilingual models) and what forms are different. against the teacher model’s scores for different at- tributes. The results are shown in Figure 1 for the 4 Submission Details training data. Perspective is a reasonable detector Attribute hs Attribute stereotype for the hate speech attribute, but performs less well 1.0 for the stereotype attribute, with the identity attack 0.8 model performing the best. Using these same models on the AMI task, True Positive Rate 0.6 shown in Figure 2 for detecting misogyny proved 0.4 even more challenging. In this case, the aggres- identity_attack 83.6% identity_attack 71.2% severe_toxicity 82.6% toxicity 82% severe_toxicity 70.9% insult 70.5% siveness attribute was evaluated only on the sub- 0.2 insult 80.8% toxicity 70.4% profanity 79% profanity 69.9% set of the training data labeled misogynous. In threat 68.1% threat 63.6% 0.0 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 this case, the most popular attribute of “toxicity” False Positive Rate False Positive Rate is actually counter-indicative of the misogyny la- bel. The best detector for both of these attributes Figure 1: ROC curves for the PerspectiveAPI appears to be the “threat” model. multilingual teacher model attributes compared to As can be seen, the existing classifiers are all the HaSpeeDe2 attributes (hate speech and stereo- poor predictors of both attributes for this shared type). task. Due to errors in our initial analysis, we did not end up using any of the models used for Per- 1.0 Attribute misogynous Attribute aggressiveness spectiveAPI in our final submissions. 1.0 0.8 n ch 0.8 ha ssio e ee yp y or sp ot i True Positive Rate bm eg 0.6 0.6 re te t Ca ste Su 0.4 0.4 news 1 0.68 0.64 threat 61.5% threat 66% identity_attack 56.5% identity_attack 61.8% 2 0.64 0.68 0.2 insult 45.2% 0.2 severe_toxicity 53.9% severe_toxicity 44.4% insult 53.3% tweets 1 0.72 0.67 toxicity 39.8% toxicity 48.4% profanity 28.8% 0.0 profanity 37% 2 0.77 0.74 0.0 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 False Positive Rate False Positive Rate Table 1: Macro-averaged F1 scores for Jigsaw’s HaSpeeDe2 Submissions. Figure 2: ROC curves for PerspectiveAPI multi- lingual teacher model attributes compared to the AMI attributes (misogyny and aggressiveness). 4.1 HaSpeeDe2 The Jigsaw team submitted two separate submis- As Jigsaw has already developed toxicity mod- sions that were independently trained for Tasks A els for the Italian language, we initially hoped and B. that these would provide a preliminary baseline for the competition despite the independent na- 4.1.1 First Submission ture of the development of the annotation guide- Our first submission, one that did not perform very lines. Our Italian models score comments for tox- well, was based on a simple multilingual BERT icity as well as five additional distinct toxicity at- model fine-tuned on 10 random splits of the train- tributes: severe toxicity, profanity, threats, insults, ing data. For each split, 10% of the data was and identity attacks. We might expect some of held out to choose an appropriate equal-error-rate these attributes to correlate with the HaSpeeDe2 threshold for the resulting model. and AMI attributes, though it is not immediately The BERT fine-tuning system used the 12 layer clear whether any of these correlations should be model (Tensorflow Hub, 2020), a batch size of particularly strong. 64 and sequence length of 128. A single dense The current Jigsaw PerspectiveAPI models are layer is used to connect to the two output sigmoids typically trained via distillation from a multilin- which are trained using a binary cross-entropy loss
using stochastic gradient descent with early stop- ping, which is computed using the AUC metric 1.0 computed using the 10% held out slice. This 0.8 model is implemented using Keras (Chollet and True Positive Rate others, 2015). 0.6 To create the final submission, the decisions of the ten separate classifiers were combined in a ma- 0.4 jority voting scheme (if 5 or more models pro- Tweets Hatespeech 85.5% 0.2 Tweets Stereotype 82.6% duced a positive detection, the attribute was as- News Stereotype 77.3% signed true). News Hatespeech 75.7% 0.0 0.0 0.2 0.4 0.6 0.8 1.0 4.1.2 Second Submission False Positive Rate Our second submission was based on a similar ap- proach of fine-tuning a BERT-based model, but Figure 3: ROC plots for HaSpeeDe2 Test Set La- one based on a more closely matched training set. bels. The underlying technology we used is the same as the Google Cloud AutoML for natural language point and custom wordpiece vocabulary from Sec- processing product that had been employed in sim- tion 4.1.2. However, a larger batch-size of 128 ilar labeling applications (Bisong, 2019). was specified. All models were fine-tuned simul- The remaining models built for this competi- taneously on misogynous and aggressive labels tion and in the subsequent section are based on a using the provided data, where zero aggressive- customized BERT 768-dimension 12-layer model ness weights were assigned to data points with no pretrained on 1B user-generated comments using misogynous labels. MLM for 125 steps. This model was then fine- tuned on supervised comments in multiple lan- Both submissions were based on ensembles of guages for six attributes: toxicity, severe toxic- partitioned models evaluated on a 10% held-out ity, obscene, threat, insult, and identity hate. This test set. We explored two different ensembling model also uses a custom wordpiece model (Wu et techniques, which we discuss in the next section. al., 2016) comprised of 200K tokens representing AMI submission 1 does not not include syn- tokens from hundreds of languages. thetic data. AMI submission 2 includes the syn- Our hate speech and misogyny models use a thetic data and custom biasing mitigation data se- fully connected final layer that combines the six lected from Wikipedia articles. Table 2 clearly output attributes and allows weight propagation shows that the inclusion of such data significantly through all layers of the network. Fine-tuning con- improved the performance on Task B for submis- tinues on using the supervised training data pro- sion 2. Interestingly, the inclusion of synthetic and vided by the competition hosts using the ADAM bias mitigation data slightly improved the perfor- optimizer with a learning rate of 1e–5. mance in Task A as well. Figure 3 displays the ROC curve for our second submission for each of the news and the tweets n Sc ssio datasets as well as for both the hate speech and i bm e k or s stereotype attributes. Su Ta Our second submission for HaSpeeDe2 con- A 1 0.738 2 0.741 sisted of fine-tuning a single model with the pro- B 1 0.649 vided training data with a 10% held-out set. The 2 0.883 custom BERT model was fine-tuned on TPUs us- Table 2: Misogynous and Aggressiveness Macro- ing a relatively small batch size of 32. averaged F1 scores for Jigsaw’s AMI Submis- 4.2 AMI sions. Our submissions for the AMI task only consid- ered the unconstrained case, due to the use of The two Jigsaw models ranked in first and sec- pretrained models. All AMI models were fine- ond place for Task A. The second submission tuned on TPUs using the customized BERT check- ranked first among participants for Task B.
4.2.1 Ensembling Models B is not surprising given that no bias mitigating Both the first and second submissions for AMI data or constraints were included in training. were ensembles of fine-tuned custom BERT mod- 4.2.3 Second Submission els constructed from partitioned training data. We explored two ensembling techniques (Brownlee, In order to mitigate bias, we decided to augment 2020): the training data set using sentences sampled from the Italian Wikipedia articles that contain the 17 • Majority Vote: Each partitioned model was terms listed in the identity terms file provided with evaluated using a model specific threshold. the test set data. These sentences were labeled The label for each attribute was determined as both non-misogynous and non-aggressive. 11K by majority vote among the models. sentences were used for this purpose, with the term frequencies summarized in Table 3. • Average: The raw models probabilities are averaged together. The combined model cal- Identity Term Sentence Count culates the labels via custom thresholds de- donna 4306 donne 3100 termined by evaluation on a held-out set. femmine 1275 femmina 652 Thresholds for the individual models in the ma- fidanzata 538 jority vote and average ensemble were calculated nonna 378 mamma 269 to optimize for the point on the held-out data ROC casalinga 256 curve where |TPR − (1 − FPR)| is minimized. casalinghe 187 The majority voting model performed slightly compagne 132 compagna 34 better for both the misogynous and aggressive task mamme 24 on the held-out sets. As such, both submissions fidanzate 12 use majority vote. nonne 11 matrone 9 4.2.2 First Submission matrona 8 morosa 6 Using the same configuration as Section 4.1.2, we partitioned the raw training data into ten randomly Table 3: Term frequency in Wikipedia sampled chosen partitions and fine-tuned nine of these us- sentences for bias mitigation. ing the 10% held out portion to compute thresh- olds. No synthetic or de-biasing data was included The second submission employed the same par- in this submission. titioning of data with a held-out set. However the We include ROC curves for half of these mod- unconstrained data included the raw training data, els in Figure 4, to illustrate that they are similar the provided synthetic data and our de-biasing but with some variance when used to score the test term data. As with submission 1, majority vote data. was used with custom thresholds determined by evaluation on the held-out set. 1.0 Our first unconstrained submission for AMI achieved scores of 0.741 for Task A and 0.883 for 0.8 Task B. True Positive Rate 0.6 0.4 5 Error Analysis misogynous-4 91.5% aggressiveness-1 89.6% 0.2 misogynous-0 90.1% misogynous-1 90% aggressiveness-3 88.5% aggressiveness-4 87.9% We discuss an informal analysis of the errors we misogynous-3 89.8% aggressiveness-2 87.9% 0.0 misogynous-2 89.3% aggressiveness-0 87.3% observed with each of these tasks. Aside from the 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 False Positive Rate False Positive Rate typical questions regarding data annotation qual- ity, and the small sample sizes, we observed some Figure 4: ROC plots for AMI test set labels for particular instances of avoidable errors. models pre-ensemble. 5.1 HaSpeeDe2 Errors Our first unconstrained submission using major- Looking at the largest incongruities as shown in ity vote for AMI achieved scores of 0.738 for Task Table 4 it is clear that context, which is unavail- A and 0.649 for Task B. The poorer score for Task able to our models, and presumably to the mod-
ID Comment HS Score 11355 @user @user @user Giustissimo, non bisogna mai nascondersi nelle ideologie, 1 .00001 sopratutto oggi perché non esistono più. Sta di fatto, che le cose più aberranti che leggi oggi sui giornali hanno sempre@a@che fare con stranieri... o rom URL 10803 #Simone di #CasalBruciato, #Roma: “Certi rom sono cittadini italiani, ma non 1 .00003 sono uguali a noi. Uguali non è il termine più giusto da usare”. URL 11288 I SOLDI DEI DISABILI AI MIGRANTI La regione Emilia Romagna destina 1 .00003 la metà dei fondi destinati alle fasce deboli a progetti per i richiedenti asilo A Reggio Emilia il 69% delle risorse stanziate sono state utilizzate ai richiedenti asilo #PRIMAGLIITALIANI URL 10420 #MeNeFottoDi questi sfigati #facciamorete che continuano a giustificare ogni 0 0.99996 crimine commesso da immigrati... non fate rete, FATE SCHIFO... #facciamo- ciFURBI 11189 @user Naturalmente in questo caso alla faccia dei comunisti e dei migranti 0 0.99996 stitici! 10483 @user SCHIFOSA IPOCRITA SPONSORIZZI I MUSSULMANI E POI VOI 0 0.99995 DARE I DIRITTI ALLE DONNE SI VEDE CHE SEI POSSEDUTA DAL DIAVOLO SEI BUGIARDA BOLDRINA SAI SOLO PROTESTARE POI TI CHIEDI PERCHÉ IL VERO ITALIANO TI ODIA PERCHÉ SEI UNA SPORCA IPOCRITA Table 4: Largest Errors for hate speech classifier on HaSpeeDe2 Tweet data erators, is important for determining the author’s amount of training data, even if it creates an imbal- intent. The use of humor and the practice of quot- ance, is one way to address this, as we did in the ing text from another author are also confounding case of the AMI challenge. factors. As this task is known to be hard (Vigna et al., 2017; van Aken et al., 2018), the edge cases 5.2 AMI display these confounding reasons. Additionally, Because we are using ensemble models trained on as evidenced by the last comment, the subtlety of partitions of the training set, we observe that the hate speech that is directed toward the designated highest-scoring test samples that are marked non- target for this challenge has not been well cap- misogynous and non-aggressive, as well as the tured. lowest-scoring misogynous and aggressive com- The BERT model that we fine-tuned for this ap- ments, vary from model to model. However, we plication is cased, and we see within our errors fre- display the most frequently occurring mistakes quent use of all-caps text. However, lower casing across all ten ensembles in Table 5. the text has almost no effect on the scores, suggest- Regarding the false alarms, these comments ap- ing that the BERT pre-training has already linked pear to be mislabeled test instances, and there is the various cased versions of the tokens in the vo- ample support for this claim in the training data. cabulary. The first comment combines both uppercase and We analyzed the frequency of word piece frag- a missing space. While it’s true that subjunctive ments in the data and saw no correlation between mode is not well represented in the training data, misclassification and the presence of segmented lower casing this sentence produces high scores. words. This suggests that vocabulary coverage in This is also the case with the third example. The the test set does not play a significant role in ex- second error seems more subtle, perhaps an at- plaining our systems’ errors. tempt at humor, but one with no salient misogyny Considering the sentence with the high model terms. score for hate speech, several single terms are 6 Bias tagged by the model. For example the term “sfi- gati” occurs only once in the training data in a Because the identity terms for AMI are not ob- sentence that is marked as non-hate speech. How- served with a high frequency in the test data, we ever, this term is not in our vocabulary and gets restrict our analysis to the synthetic data set. We split into pieces “sfiga##ti”, and the prefix “sfiga” find wide variation in the performance of our in- appears in two out of three training examples that dividual models, with one model exhibiting very are marked hate speech—exactly the kind of data poor performance across the subgroups. The sum- sparsity that leads to unwanted bias. Using a larger mary of the AUC measurements for these models
ID Comment Error Type 5466 Odio Sakura per il semplice motivo che qualunque cosa faccia o dica Naruto Missed Detection lei lo prende a schiaffi o a pugniHA CHIESTO COME STA SAI DIOSANTO BRUTTA STRONZA MA CONTRALLI MADONNA SPERO CHE TI UCCI- DANOscusami Sarada 5385 porca troia la prenderei a schiaffi se non fosse mia madre Missed Detection 5819 ma in tutto ciò lo scopo di anna qual è? far soffrire il mio protetto? IO TI Missed Detection AMMAZZO COI LANCIAFIAMME OH #TemptationIsland 5471 @danielita8811 Che bel culo tutto da sfondare False Alarm 5604 @coppiacalda2 Che bel culo da inculare False Alarm Table 5: Persistent errors for AMI across different ensembles. are shown in Figure 5, Figure 6, and Figure 7 using the technique presented in (Borkan et al., 2019). There does not appear to be a systemic problem 0 1 2 3 4 5 6 7 8 9 1.0 nonne 0.95 0.95 0.90 0.82 0.87 0.22 0.83 0.83 0.92 0.73 with bias in these models, but judging based only matrone 0.96 0.97 0.93 0.94 0.92 0.30 0.94 0.96 0.95 0.95 mamme 0.95 0.94 0.94 0.94 0.92 0.15 0.92 0.94 0.95 0.90 upon synthetic data is probably unwise. The single casalinghe 0.97 0.97 0.94 0.95 0.94 0.26 0.93 0.95 0.95 0.95 0.8 term “donna” from the test set shows a subgroup compagne 0.97 0.96 0.93 0.94 0.95 0.22 0.92 0.95 0.95 0.95 morose 0.96 0.92 0.88 0.87 0.90 0.44 0.88 0.85 0.91 0.87 AUC that drops substantially from the background 0.98 0.97 0.97 0.95 0.96 0.28 0.96 0.95 0.96 0.95 femmine AUC for nearly all of the models, perhaps indicat- donne 0.98 0.96 0.96 0.94 0.94 0.45 0.96 0.96 0.94 0.96 0.6 fidanzate 0.97 0.97 0.95 0.93 0.93 0.24 0.92 0.93 0.96 0.94 ing limitations of judging based on synthetic data. 0.96 0.97 0.95 0.94 0.95 0.69 0.95 0.95 0.95 0.93 nonna matrona 0.95 0.96 0.95 0.95 0.94 0.76 0.94 0.95 0.95 0.94 0.4 casalinga 0.96 0.97 0.96 0.95 0.95 0.50 0.95 0.95 0.95 0.94 morosa 0.96 0.97 0.95 0.94 0.93 0.70 0.93 0.93 0.95 0.93 femmina 0.99 0.98 0.95 0.95 0.96 0.55 0.96 0.96 0.96 0.95 0 1 2 3 4 5 6 7 8 9 1.0 nonne 0.95 0.90 0.81 0.59 0.68 0.44 0.62 0.71 0.61 0.55 0.2 mamma 0.96 0.97 0.95 0.94 0.95 0.68 0.95 0.95 0.94 0.92 matrone 0.97 0.96 0.93 0.94 0.93 0.43 0.94 0.95 0.96 0.94 donna 0.98 0.97 0.96 0.96 0.97 0.70 0.98 0.96 0.95 0.96 mamme 0.96 0.96 0.96 0.94 0.93 0.46 0.91 0.94 0.95 0.93 fidanzata 0.96 0.95 0.93 0.92 0.93 0.76 0.93 0.94 0.94 0.93 casalinghe 0.97 0.97 0.95 0.96 0.96 0.50 0.94 0.95 0.96 0.95 0.8 compagna 0.96 0.97 0.95 0.95 0.95 0.64 0.94 0.95 0.95 0.95 compagne 0.96 0.96 0.93 0.93 0.94 0.47 0.92 0.94 0.96 0.94 0.0 morose 0.96 0.93 0.89 0.90 0.92 0.49 0.89 0.86 0.92 0.92 femmine 0.97 0.95 0.94 0.95 0.93 0.45 0.94 0.94 0.96 0.96 0.97 0.96 0.94 0.95 0.94 0.47 0.95 0.94 0.94 0.95 0.6 donne fidanzate 0.98 0.97 0.95 0.95 0.94 0.41 0.92 0.94 0.97 0.92 nonna 0.97 0.98 0.96 0.96 0.96 0.44 0.97 0.96 0.96 0.94 Figure 6: Background Positive, Subgroup Nega- matrona 0.96 0.97 0.95 0.97 0.95 0.52 0.96 0.96 0.97 0.96 0.4 casalinga 0.96 0.97 0.95 0.97 0.97 0.49 0.97 0.95 0.96 0.95 tive AUC morosa 0.97 0.96 0.95 0.95 0.95 0.45 0.94 0.95 0.95 0.96 femmina 0.97 0.98 0.96 0.96 0.96 0.48 0.96 0.96 0.96 0.96 mamma 0.96 0.97 0.96 0.96 0.95 0.49 0.96 0.94 0.95 0.94 0.2 donna 0.97 0.97 0.95 0.96 0.97 0.45 0.97 0.94 0.96 0.93 fidanzata 0.97 0.96 0.95 0.95 0.96 0.48 0.95 0.95 0.96 0.95 compagna 0.96 0.97 0.95 0.96 0.95 0.47 0.96 0.95 0.96 0.95 0 1 2 3 4 5 6 7 8 9 0.0 1.0 nonne 0.97 0.94 0.92 0.77 0.86 0.73 0.87 0.92 0.81 0.83 matrone 0.97 0.95 0.95 0.93 0.94 0.62 0.93 0.92 0.95 0.92 mamme 0.97 0.97 0.95 0.93 0.94 0.80 0.93 0.93 0.95 0.95 casalinghe 0.96 0.96 0.95 0.94 0.95 0.71 0.94 0.94 0.96 0.92 0.8 compagne 0.96 0.96 0.94 0.93 0.93 0.73 0.94 0.92 0.95 0.91 Figure 5: Subgroup AUC morose 0.96 0.97 0.94 0.95 0.95 0.53 0.94 0.95 0.95 0.96 femmine 0.95 0.94 0.92 0.92 0.90 0.65 0.91 0.92 0.93 0.93 0.96 0.96 0.93 0.94 0.93 0.50 0.91 0.92 0.95 0.91 0.6 donne fidanzate 0.97 0.96 0.94 0.95 0.95 0.66 0.93 0.94 0.95 0.91 nonna 0.97 0.97 0.96 0.96 0.96 0.26 0.96 0.95 0.97 0.94 7 Conclusions and Future Work matrona 0.98 0.97 0.94 0.94 0.95 0.21 0.95 0.94 0.96 0.95 0.4 casalinga 0.97 0.97 0.94 0.95 0.95 0.47 0.96 0.95 0.96 0.93 0.97 0.96 0.94 0.94 0.95 0.23 0.94 0.95 0.95 0.95 Both of these challenges dealt with issues re- morosa femmina 0.94 0.96 0.94 0.93 0.93 0.40 0.92 0.94 0.94 0.94 lated to content moderation and evaluation of user- mamma 0.97 0.97 0.95 0.95 0.94 0.28 0.95 0.93 0.95 0.94 0.2 0.95 0.96 0.93 0.93 0.93 0.25 0.92 0.92 0.95 0.89 generated content. While early research raised donna fidanzata 0.97 0.97 0.96 0.97 0.96 0.20 0.96 0.96 0.97 0.95 fears of censorship, the ongoing challenges plat- compagna 0.97 0.97 0.94 0.94 0.94 0.33 0.95 0.94 0.97 0.93 0.0 forms face have made it necessary to consider the potential of machine learning. Advances in natu- ral language understanding have produced models that work surprisingly well, even ones that are able Figure 7: Background Negative, Subgroup Posi- to detect malicious intent that users try to encode tive AUC in subtle ways.
Our particular approach to the EVALITA chal- Jason Brownlee. 2020. How to develop vot- lenges represented an unsurprising application of ing ensembles with python. https:// machinelearningmastery.com/voting- what has now become a textbook technique: lever- ensembles-with-python/, September. aging the resources of large pre-trained models. However, many participants achieved nearly simi- Francois Chollet et al. 2015. Keras. lar performance levels in the constrained task. We Jacob Devlin, Ming-Wei Chang, Kenton Lee, and regard this as a more impressive accomplishment. Kristina Toutanova. 2019. BERT: Pre-training of Jigsaw continues to apply machine learning to deep bidirectional transformers for language under- support publishers and to help them host quality standing. In Proceedings of the 2019 Conference of online conversations where readers feel safe par- the North American Chapter of the Association for Computational Linguistics: Human Language Tech- ticipating. The kinds of comments these chal- nologies, Volume 1 (Long and Short Papers), pages lenges tagged are some of the most concerning 4171–4186, Minneapolis, Minnesota, June. Associ- and pernicious online behaviors, far outside of the ation for Computational Linguistics. norms that are tolerated in other public spaces. Paolo Rosso Elisabetta Fersini, Debora Nozza. 2020. But humans and machines both still misinterpret Ami @ evalita2020: Automatic misogyny identi- profanity for hostility, and tagging humor, quo- fication. In Valerio Basile, Danilo Croce, Maria tations, sarcasm, and other legitimate expressions Di Maro, and Lucia C. Passaro, editors, Proceedings for moderation remain serious problems. of the 7th evaluation campaign of Natural Language Processing and Speech tools for Italian (EVALITA Challenges like the AMI and HasSpeede2 com- 2020), Online. CEUR.org. petitions underscore the importance of under- standing the relationships between the parties in a Suchin Gururangan, Ana Marasović, Swabha conversation, and the participants’ intents. We are Swayamdipta, Kyle Lo, Iz Beltagy, Doug Downey, and Noah A. Smith. 2020. Don’t stop pretraining: greatly encouraged that attributes that our systems Adapt language models to domains and tasks. do not currently capture were somewhat within the In Proceedings of the 58th Annual Meeting of reach of our present techniques—but clearly much the Association for Computational Linguistics, work remains to be done. pages 8342–8360, Online, July. Association for Computational Linguistics. Jigsaw. 2018. Jigsaw toxic comment classifi- References cation challenge. https://www.kaggle. Valerio Basile, Danilo Croce, Maria Di Maro, and Lu- com/c/jigsaw-toxic-comment- cia C. Passaro. 2020. Evalita 2020: Overview classification-challenge, March. of the 7th evaluation campaign of natural language processing and speech tools for italian. In Valerio Jigsaw. 2019. Jigsaw unintended bias in toxic- Basile, Danilo Croce, Maria Di Maro, and Lucia C. ity classification. https://www.kaggle. Passaro, editors, Proceedings of Seventh Evalua- com/c/jigsaw-unintended-bias-in- tion Campaign of Natural Language Processing and toxicity-classification, July. Speech Tools for Italian. Final Workshop (EVALITA Jigsaw. 2020. Jigsaw multilingual toxic com- 2020), Online. CEUR.org. ment classification. https://www.kaggle. Ekaba Bisong, 2019. Google AutoML: Cloud Natu- com/c/jigsaw-multilingual-toxic- ral Language Processing, pages 599–612. Apress, comment-classification, July. Berkeley, CA. Endang Wahyu Pamungkas, Valerio Basile, and Vi- Daniel Borkan, Lucas Dixon, Jeffrey Sorensen, Nithum viana Patti. 2020. Misogyny detection in twitter: Thain, and Lucy Vasserman. 2019. Nuanced met- a multilingual and cross-domain study. Information rics for measuring unintended bias with real data Processing & Management, 57(6):102360. for text classification. In Companion Proceedings of The 2019 World Wide Web Conference, pages 491– Telmo Pires, Eva Schlinger, and Dan Garrette. 2019. 500. How multilingual is multilingual BERT? In Pro- ceedings of the 57th Annual Meeting of the Asso- Cristina Bosco, Tommaso Caselli, Gloria Comandini, ciation for Computational Linguistics, pages 4996– Elisa Di Nuovo, Simona Frenda, Viviana Patti, 5001, Florence, Italy, July. Association for Compu- Irene Russo, Manuela Sanguinetti, and Marco tational Linguistics. Stranisci. 2020. Hate speech detection task second edition (haspeede2) at evalita 2020 task guidelines. Tensorflow Hub. 2020. Mulilingual L12 H768 A12 https://github.com/msang/haspeede/ V2. https://tfhub.dev/tensorflow/ blob/master/2020/HaSpeeDe2020_ bert_multi_cased_L-12_H-768_A-12/2, Task_guidelines.pdf. August.
Betty van Aken, Julian Risch, Ralf Krestel, and Alexander Löser. 2018. Challenges for toxic com- ment classification: An in-depth error analysis. In Proceedings of the 2nd Workshop on Abusive Lan- guage Online (ALW2), pages 33–42, Brussels, Bel- gium, October. Association for Computational Lin- guistics. F. D. Vigna, A. Cimino, Felice Dell’Orletta, M. Petroc- chi, and M. Tesconi. 2017. Hate me, hate me not: Hate speech detection on facebook. In ITASEC. Shijie Wu and Mark Dredze. 2019. Beto, bentz, be- cas: The surprising cross-lingual effectiveness of BERT. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natu- ral Language Processing (EMNLP-IJCNLP), pages 833–844, Hong Kong, China, November. Associa- tion for Computational Linguistics. Yonghui Wu, Mike Schuster, Zhifeng Chen, Quoc V. Le, Mohammad Norouzi, Wolfgang Macherey, Maxim Krikun, Yuan Cao, Qin Gao, Klaus Macherey, Jeff Klingner, Apurva Shah, Melvin Johnson, Xiaobing Liu, Lukasz Kaiser, Stephan Gouws, Yoshikiyo Kato, Taku Kudo, Hideto Kazawa, Keith Stevens, George Kurian, Nishant Patil, Wei Wang, Cliff Young, Jason Smith, Jason Riesa, Alex Rudnick, Oriol Vinyals, Greg Corrado, Macduff Hughes, and Jeffrey Dean. 2016. Google’s neural machine translation system: Bridging the gap between human and machine translation. CoRR, abs/1609.08144.
You can also read