A Survey of Data Augmentation Approaches for NLP - ACL ...
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
A Survey of Data Augmentation Approaches for NLP Steven Y. Feng∗, 1 Varun Gangal∗, 1 Jason Wei†, 2 Sarath Chandar,3 Soroush Vosoughi,4 Teruko Mitamura,1 Eduard Hovy1 1 Carnegie Mellon University, 2 Google Research 3 Mila - Quebec AI Institute, 4 Dartmouth College {syfeng,vgangal,teruko,hovy}@cs.cmu.edu jasonwei@google.com sarath.chandar@mila.quebec soroush@dartmouth.edu Abstract Data augmentation has recently seen increased interest in NLP due to more work in low- resource domains, new tasks, and the popu- larity of large-scale neural networks that re- quire large amounts of training data. De- spite this recent upsurge, this area is still rel- atively underexplored, perhaps due to the chal- lenges posed by the discrete nature of language data. In this paper, we present a comprehen- sive and unifying survey of data augmenta- tion for NLP by summarizing the literature in Figure 1: Weekly Google Trends scores for the search a structured manner. We first introduce and term "data augmentation", with a control, uneventful motivate data augmentation for NLP, and then ML search term ("minibatch") for comparison. discuss major methodologically representative approaches. Next, we highlight techniques Despite these challenges, there has been in- that are used for popular NLP applications and creased interest and demand for DA for NLP. As tasks. We conclude by outlining current chal- NLP grows due to off-the-shelf availability of large lenges and directions for future research. Over- pretrained models, there are increasingly more all, our paper aims to clarify the landscape of existing literature in data augmentation for tasks and domains to explore. Many of these are NLP and motivate additional work in this area. low-resource, and have a paucity of training exam- We also present a GitHub repository with a pa- ples, creating many use-cases for which DA can per list that will be continuously updated at play an important role. Particularly, for many non- https://github.com/styfeng/DataAug4NLP. classification NLP tasks such as span-based tasks and generation, DA research is relatively sparse 1 Introduction despite their ubiquity in real-world settings. Data augmentation (DA) refers to strategies for in- Our paper aims to sensitize the NLP community creasing the diversity of training examples without towards this growing area of work, which has also explicitly collecting new data. It has received active seen increasing interest in ML overall (as seen in attention in recent machine learning (ML) research Figure 1). As interest and work on this topic con- in the form of well-received, general-purpose tech- tinue to increase, this is an opportune time for a niques such as UDA (Xie et al., 2020) (3.1) and paper of our kind to (i) give a bird’s eye view of M IX U P (Zhang et al., 2017) (3.2). These are often DA for NLP, and (ii) identify key challenges to first explored in computer vision (CV), and DA’s effectively motivate and orient interest in this area. adaptation for natural language processing (NLP) To the best of our knowledge, this is the first survey seems secondary and comparatively underexplored, to take a detailed look at DA methods for NLP.1 perhaps due to challenges presented by the discrete This paper is structured as follows. Section nature of language, which rules out continuous 1 noising and makes it hard to maintain invariance. Liu et al. (2020a) present a smaller-scale text data aug- mentation survey that is concise and focused. Our work serves ∗ Equal contribution by the two authors. as a more comprehensive survey with larger coverage and is † AI Resident. more up-to-date. 968 Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 968–988 August 1–6, 2021. ©2021 Association for Computational Linguistics
2 discusses what DA is, its goals and trade-offs, Rule-based techniques are easy-to-implement and why it works. Section 3 describes popular but usually offer incremental performance improve- methodologically representative DA techniques for ments (Li et al., 2017; Wei and Zou, 2019; Wei NLP—which we categorize into rule-based (3.1), et al., 2021b). Techniques leveraging trained mod- example interpolation-based (3.2), or model-based els may be more costly to implement but introduce (3.3). Section 4 discusses useful NLP applications more data variation, leading to better performance for DA, including low-resource languages (4.1), boosts. Model-based techniques customized for mitigating bias (4.2), fixing class imbalance (4.3), downstream tasks can have strong effects on per- few-shot learning (4.4), and adversarial examples formance but be difficult to develop and utilize. (4.5). Section 5 describes DA methods for common Further, the distribution of augmented data NLP tasks including summarization (5.1), question should neither be too similar nor too different from answering (5.2), sequence tagging tasks (5.3), pars- the original. This may lead to greater overfitting ing tasks (5.4), grammatical error correction (5.5), or poor performance through training on examples neural machine translation (5.6), data-to-text NLG not representative of the given domain, respectively. (5.7), open-ended and conditional text generation Effective DA approaches should aim for a balance. (5.8), dialogue (5.9), and multimodal tasks (5.10). Kashefi and Hwa (2020) devise a KL- Finally, Section 6 discusses challenges and future Divergence-based unsupervised procedure to pre- directions in DA for NLP. Appendix A lists useful emptively choose among DA heuristics, rather than blog posts and code repositories. a typical "run-all-heuristics" comparison, which Through this work, we hope to emulate past pa- can be very time and cost intensive. pers which have surveyed DA methods for other types of data, such as images (Shorten and Khosh- Interpretation of DA Dao et al. (2019) note that goftaar, 2019), faces (Wang et al., 2019b), and time "data augmentation is typically performed in an ad- series (Iwana and Uchida, 2020). We hope to draw hoc manner with little understanding of the under- further attention, elicit broader interest, and moti- lying theoretical principles", and claim the typical vate additional work in DA, particularly for NLP. explanation of DA as regularization to be insuffi- cient. Overall, there indeed appears to be a lack of 2 Background research on why exactly DA works. Existing work on this topic is mainly surface-level, and rarely What is data augmentation? Data augmentation investigates the theoretical underpinnings and prin- (DA) encompasses methods of increasing training ciples. We discuss this challenge more in §6, and data diversity without directly collecting more data. highlight some of the existing work below. Most strategies either add slightly modified copies Bishop (1995) show training with noised exam- of existing data or create synthetic data, aiming for ples is reducible to Tikhonov regularization (sub- the augmented data to act as a regularizer and re- sumes L2). Rajput et al. (2019) show that DA can duce overfitting when training ML models (Shorten increase the positive margin for classifiers, but only and Khoshgoftaar, 2019; Hernández-García and when augmenting exponentially many examples König, 2020). DA has been commonly used in for common DA methods. CV, where techniques like cropping, flipping, and color jittering are a standard component of model Dao et al. (2019) think of DA transformations training. In NLP, where the input space is discrete, as kernels, and find two ways DA helps: averaging how to generate effective augmented examples that of features and variance regularization. Chen et al. capture the desired invariances is less obvious. (2020d) show that DA leads to variance reduction by averaging over orbits of the group that keep the What are the goals and trade-offs? Despite data distribution approximately invariant. challenges associated with text, many DA tech- niques for NLP have been proposed, ranging from 3 Techniques & Methods rule-based manipulations (Zhang et al., 2015) to more complicated generative approaches (Liu et al., We now discuss some methodologically represen- 2020b). As DA aims to provide an alternative to tative DA techniques which are relevant to all tasks collecting more data, an ideal DA technique should via the extensibility of their formulation.2 be both easy-to-implement and improve model per- 2 Table 1 compares several DA methods by various aspects formance. Most offer trade-offs between these two. relating to their applicability, dependencies, and requirements. 969
3.2 Example Interpolation Techniques Another class of DA techniques, pioneered by M IX U P (Zhang et al., 2017), interpolates the in- puts and labels of two or more real examples. This class of techniques is also sometimes referred to as Mixed Sample Data Augmentation (MSDA). Ensu- ing work has explored interpolating inner compo- nents (Verma et al., 2019; Faramarzi et al., 2020), more general mixing schemes (Guo, 2020), and adding adversaries (Beckham et al., 2019). Another class of extensions of M IX U P which has been growing in the vision community attempts to fuse raw input image pairs together into a single input image, rather than improve the continuous in- Figure 2: Dependency tree morphing DA applied to a terpolation mechanism. Examples of this paradigm Turkish sentence, Şahin and Steedman (2018) include C UT M IX (Yun et al., 2019), C UT O UT (De- Vries and Taylor, 2017) and C OPY-PASTE (Ghiasi 3.1 Rule-Based Techniques et al., 2020). For instance, C UT M IX replaces a small sub-region of Image A with a patch sampled Here, we cover DA primitives which use easy- from Image B, with the labels mixed in proportion to-compute, predetermined transforms sans model to sub-region sizes. There is potential to borrow components. Feature space DA approaches gen- ideas and inspiration from these works for NLP, erate augmented examples in the model’s feature e.g. for multimodal work involving both images space rather than input data. Many few-shot learn- and text (see "Multimodal challenges" in §6). ing approaches (Hariharan and Girshick, 2017; A bottleneck to using M IX U P for NLP tasks Schwartz et al., 2018) leverage estimated feature was the requirement of continuous inputs. This has space "analogy" transformations between exam- been overcome by mixing embeddings or higher ples of known classes to augment for novel classes hidden layers (Chen et al., 2020c). Later variants (see §4.4). Paschali et al. (2019) use iterative propose speech-tailored mixing schemes (Jindal affine transformations and projections to maximally et al., 2020b) and interpolation with adversarial "stretch" an example along the class-manifold. examples (Cheng et al., 2020), among others. Wei and Zou (2019) propose E ASY DATA AUG - S EQ 2M IX U P (Guo et al., 2020) generalizes MENTATION (EDA), a set of token-level random M IX U P for sequence transduction tasks in two perturbation operations including random insertion, ways - the "hard" version samples a binary mask deletion, and swap. They show improved perfor- (from a Bernoulli with a β(α, α) prior) and picks mance on many text classification tasks. UDA (Xie from one of two sequences at each token position, et al., 2020) show how supervised DA methods can while the "soft" version softly interpolates between be exploited for unsupervised data through consis- sequences based on a coefficient sampled from tency training on (x, DA(x)) pairs. β(α, α). The "soft" version is found to outperform For paraphrase identification, Chen et al. (2020b) the "hard" version and earlier interpolation-based construct a signed graph over the data, with indi- techniques like S WITCH O UT (Wang et al., 2018a). vidual sentences as nodes and pair labels as signed 3.3 Model-Based Techniques edges. They use balance theory and transitivity to infer augmented sentence pairs from this graph. Seq2seq and language models have also been used Motivated by image cropping and rotation, Şahin for DA. The popular BACKTRANSLATION method and Steedman (2018) propose dependency tree mor- (Sennrich et al., 2016) translates a sequence into phing. For dependency-annotated sentences, chil- another language and then back into the original dren of the same parent are swapped (à la rotation) language. Kumar et al. (2019a) train seq2seq mod- or some deleted (à la cropping), as seen in Figure 2. els with their proposed method DiPS which learns This is most beneficial for language families with to generate diverse paraphrases of input text using rich case marking systems (e.g. Baltic and Slavic). a modified decoder with a submodular objective, 970
from the training set and sample from it. Anaby- Tavor et al. (2020) learn a label-conditioned gen- erator by finetuning GPT-2 (Radford et al., 2019) on the training data, using this to generate candi- date examples per class. A classifier trained on the original training set is then used to select top k can- didate examples which confidently belong to the respective class for augmentation. Quteineh et al. (2020) use a similar label-conditioned GPT-2 gen- eration method, and demonstrate its effectiveness as a DA method in an active learning setup. Other approaches include syntactic or controlled Figure 3: Contextual Augmentation, Kobayashi (2018) paraphrasing (Iyyer et al., 2018; Kumar et al., 2020), document or story-level paraphrasing (Gan- and show its effectiveness as DA for several classi- gal et al., 2021), augmenting misclassified exam- fication tasks. Pretrained language models such as ples (Dreossi et al., 2018), BERT cross-encoder RNNs (Kobayashi, 2018) and transformers (Yang labeling of new inputs (Thakur et al., 2021), and et al., 2020) have also been used for augmentation. guided generation using large-scale generative lan- Kobayashi (2018) generate augmented examples guage models (Liu et al., 2020b,c). Models can by replacing words with others randomly drawn also learn to combine together simpler DA primi- according to the recurrent language model’s dis- tives (Cubuk et al., 2018; Ratner et al., 2017) or add tribution based on the current context (illustra- human-in-the-loop (Kaushik et al., 2020, 2021). tion in Figure 3). Yang et al. (2020) propose G- DAUGc which generates synthetic examples using 4 Applications pretrained transformer language models, and se- lects the most informative and diverse set for aug- In this section, we discuss several DA methods for mentation. Gao et al. (2019) advocate retaining the some common NLP applications.2 full distribution through "soft" augmented exam- ples, showing gains on machine translation. 4.1 Low-Resource Languages Nie et al. (2020) augment word representations Low-resource languages are an important and chal- with a context-sensitive attention-based mixture of lenging application for DA, typically for neural their semantic neighbors from a pretrained embed- machine translation (NMT). Techniques using ex- ding space, and show its effectiveness for NER ternal knowledge such as WordNet (Miller, 1995) on social media text. Inspired by denoising au- may be difficult to use effectively here.3 There toencoders, Ng et al. (2020) use a corrupt-and- are ways to leverage high-resource languages for reconstruct approach, with the corruption function low-resource languages, particularly if they have q(x0 |x) masking an arbitrary number of word po- similar linguistic properties. Xia et al. (2019) use sitions and the reconstruction function r(x|x0 ) un- this approach to improve low-resource NMT. masking them using BERT (Devlin et al., 2019). Li et al. (2020b) use backtranslation and self- Their approach works well on domain-shifted test learning to generate augmented training data. In- sets across 9 datasets on sentiment, NLI, and NMT. spired by work in CV, Fadaee et al. (2017) gener- Feng et al. (2019) propose a task called S EMAN - ate additional training examples that contain low- TIC T EXT E XCHANGE (STE) which involves ad- frequency (rare) words in synthetically created con- justing the overall semantics of a text to fit the texts. Qin et al. (2020) present a DA framework to context of a new word/phrase that is inserted called generate multi-lingual code-switching data to fine- the replacement entity (RE). They do so by using a tune multilingual-BERT. It encourages the align- system called SMERTI and a masked LM approach. ment of representations from source and multiple While not proposed directly for DA, it can be used target languages once by mixing their context in- as such, as investigated in Feng et al. (2020). formation. They see improved performance across Rather than starting from an existing exam- 5 tasks with 19 languages. ple and modifying it, some model-based DA ap- 3 proaches directly estimate a generative process Low-resource language challenges discussed more in §6. 971
DA Method Ext.Know Pretrained Preprocess Level Task-Agnostic S YNONYM R EPLACEMENT (Zhang et al., 2015) 3 × tok Input 3 R ANDOM D ELETION (Wei and Zou, 2019) × × tok Input 3 R ANDOM S WAP (Wei and Zou, 2019) × × tok Input 3 BACKTRANSLATION (Sennrich et al., 2016) × 3 Depends Input 3 SCPN (Wieting and Gimpel, 2017) × 3 const Input 3 S EMANTIC T EXT E XCHANGE (Feng et al., 2019) × 3 const Input 3 C ONTEXTUAL AUG (Kobayashi, 2018) × 3 - Input 3 LAMBADA (Anaby-Tavor et al., 2020) × 3 - Input × GECA (Andreas, 2020) × × tok Input × S EQ M IX U P (Guo et al., 2020) × × tok Input × S WITCH O UT (Wang et al., 2018b) × × tok Input × E MIX (Jindal et al., 2020a) × × - Emb/Hidden 3 S PEECH M IX (Jindal et al., 2020b) × × - Emb/Hidden Speech/Audio M IX T EXT (Chen et al., 2020c) × × - Emb/Hidden 3 S IGNED G RAPH (Chen et al., 2020b) × × - Input × DT REE M ORPH (Şahin and Steedman, 2018) × × dep Input 3 Sub2 (Shi et al., 2021) × × dep Input Substructural DAGA (Ding et al., 2020) × × tok Input+Label × WN-H YPERS (Feng et al., 2020) 3 × const+KWE Input 3 S YNTHETIC N OISE (Feng et al., 2020) × × tok Input 3 UE DIN -MS (DA part) (Grundkiewicz et al., 2019) 3 × tok Input 3 N ONCE (Gulordava et al., 2018) 3 × const Input 3 XLDA (Singh et al., 2019) × 3 Depends Input 3 S EQ M IX (Zhang et al., 2020) × 3 tok Input+Label × S LOT-S UB -LM (Louvan and Magnini, 2020) × 3 tok Input 3 UBT & TBT (Vaibhav et al., 2019) × 3 Depends Input 3 S OFT C ONTEXTUAL DA (Gao et al., 2019) × 3 tok Emb/Hidden 3 DATA D IVERSIFICATION (Nguyen et al., 2020) × 3 Depends Input 3 D I PS (Kumar et al., 2019a) × 3 tok Input 3 AUGMENTED SBERT (Thakur et al., 2021) × 3 - Input+Label Sentence Pairs Table 1: Comparing a selection of DA methods by various aspects relating to their applicability, dependencies, and requirements. Ext.Know, KWE, tok, const, and dep stand for External Knowledge, keyword extraction, tokeniza- tion, constituency parsing, and dependency parsing, respectively. Ext.Know refers to whether the DA method re- quires external knowledge (e.g. WordNet) and Pretrained if it requires a pretrained model (e.g. BERT). Preprocess denotes preprocessing required, Level denotes the depth at which data is modified by the DA, and Task-Agnostic refers to whether the DA method can be applied to different tasks. See Appendix B for further explanation. 4.2 Mitigating Bias THETIC M INORITY OVERSAMPLING T ECHNIQUE Zhao et al. (2018) attempt to mitigate gender (SMOTE) (Chawla et al., 2002), which gener- bias in coreference resolution by creating an aug- ates augmented minority class examples through mented dataset identical to the original but biased interpolation, still remains popular (Fernández towards the underrepresented gender (using gen- et al., 2018). M ULTILABEL SMOTE (MLSMOTE) der swapping of entities such as replacing "he" (Charte et al., 2015) modifies SMOTE to balance with "she") and train on the union of the two classes for multi-label classification, where classi- datasets. Lu et al. (2020) formally propose COUN - fiers predict more than one class at the same time. TERFACTUAL DA (CDA) for gender bias mitiga- Other techniques such as EDA (Wei and Zou, 2019) tion, which involves causal interventions that break can possibly be used for oversampling as well. associations between gendered and gender-neutral words. Zmigrod et al. (2019) and Hall Maudslay 4.4 Few-Shot Learning et al. (2019) propose further improvements to CDA. DA methods can ease few-shot learning by adding Moosavi et al. (2020) augment training sentences more examples for novel classes introduced in the with their corresponding predicate-argument struc- few-shot phase. Hariharan and Girshick (2017) tures, improving the robustness of transformer mod- use learned analogy transformations φ(z1 , z2 , x) els against various types of biases. between example pairs from a non-novel class z1 → z2 to generate augmented examples x → x0 4.3 Fixing Class Imbalance for novel classes. Schwartz et al. (2018) generalize Fixing class imbalance typically involves a combi- this to beyond just linear offsets, through their "∆- nation of undersampling and oversampling. S YN - network" autoencoder which learns the distribution 972
P (z2 |z1 , C) from all yz∗1 = yz∗2 = C pairs, where 5.2 Question Answering (QA) C is a class and y is the ground-truth labelling Longpre et al. (2019) investigate various DA and function. Both these methods are applied only on sampling techniques for domain-agnostic QA in- image tasks, but their theoretical formulations are cluding paraphrasing by backtranslation. Yang generally applicable, and hence we discuss them. et al. (2019) propose a DA method using distant Kumar et al. (2019b) apply these and other supervision to improve BERT finetuning for open- DA methods for few-shot learning of novel intent domain QA. Riabi et al. (2020) leverage Question classes in task-oriented dialog. Wei et al. (2021a) Generation models to produce augmented exam- show that data augmentation facilitates curriculum ples for zero-shot cross-lingual QA. Singh et al. learning for training triplet networks for few-shot (2019) propose XLDA, or CROSS - LINGUAL DA, text classification. Lee et al. (2021) use T5 to gen- which substitutes a portion of the input text with erate additional examples for data-scarce classes. its translation in another language, improving per- formance across multiple languages on NLI tasks 4.5 Adversarial Examples (AVEs) including the SQuAD QA task. Asai and Hajishirzi Adversarial examples can be generated using (2020) use logical and linguistic knowledge to gen- innocuous label-preserving transformations (e.g. erate additional training data to improve the accu- paraphrasing) that fool state-of-the-art NLP mod- racy and consistency of QA responses by models. els, as shown in Jia et al. (2019). Specifically, Yu et al. (2018) introduce a new QA architecture they add sentences with distractor spans to pas- called QANet that shows improved performance sages to construct AVEs for span-based QA. Zhang on SQuAD when combined with augmented data et al. (2019d) construct AVEs for paraphrase de- generated using backtranslation. tection using word swapping. Kang et al. (2018) and Glockner et al. (2018) create AVEs for textual 5.3 Sequence Tagging Tasks entailment using WordNet relations. Ding et al. (2020) propose DAGA, a two-step DA process. First, a language model over sequences of 5 Tasks tags and words linearized as per a certain scheme is In this section, we discuss several DA works learned. Second, sequences are sampled from this for common NLP tasks.2 We focus on non- language model and de-linearized to generate new classification tasks as classification is worked on examples. Şahin and Steedman (2018), discussed by default, and well covered in earlier sections (e.g. in §3.1, use dependency tree morphing (Figure 2) §3 and §4). Numerous previously mentioned DA to generate additional training examples on the techniques, e.g. (Wei and Zou, 2019; Chen et al., downstream task of part-of-speech (POS) tagging. 2020b; Anaby-Tavor et al., 2020), have been used Dai and Adel (2020) modify DA techniques pro- or can be used for text classification tasks. posed for sentence-level tasks for named entity recognition (NER), including label-wise token and 5.1 Summarization synonym replacement, and show improved perfor- Fabbri et al. (2020) investigate backtranslation as a mance using both recurrent and transformer models. DA method for few-shot abstractive summarization Zhang et al. (2020) propose a DA method based with the use of a consistency loss inspired by UDA. on M IX U P called S EQ M IX for active sequence la- Parida and Motlicek (2019) propose an iterative DA beling by augmenting queried samples, showing approach for abstractive summarization that uses a improvements on NER and Event Detection. mix of synthetic and real data, where the former is generated from Common Crawl. Zhu et al. (2019) 5.4 Parsing Tasks introduce a query-focused summarization (Dang, Jia and Liang (2016) propose DATA RECOMBINA - 2005) dataset collected using Wikipedia called TION for injecting task-specific priors to neural se- W IKI R EF which can be used for DA. Pasunuru et al. mantic parsers. A synchronous context-free gram- (2021) use DA methods to construct two training mar (SCFG) is induced from training data, and datasets for Query-focused Multi-Document Sum- new "recombinant" examples are sampled. Yu et al. marization (QMDS) called Q MDS C NN and Q MD - (2020) introduce G RAPPA, a pretraining approach S I R by modifying CNN/DM (Hermann et al., 2015) for table semantic parsing, and generate synthetic and mining search-query logs, respectively. question-SQL pairs via an SCFG. Andreas (2020) 973
use compositionality to construct synthetic exam- related words over the vocabulary. Nguyen et al. ples for downstream tasks like semantic parsing. (2020) propose DATA D IVERSIFICATION which Fragments of original examples are replaced with merges original training data with the predictions fragments from other examples in similar contexts. of several forward and backward models. Vania et al. (2019) investigate DA for low- resource dependency parsing including dependency 5.7 Data-to-Text NLG tree morphing from Şahin and Steedman (2018) Data-to-text NLG refers to tasks which require gen- (Figure 2) and modified nonce sentence genera- erating natural language descriptions of structured tion from Gulordava et al. (2018), which replaces or semi-structured data inputs, e.g. game score content words with other words of the same POS, tables (Wiseman et al., 2017). Randomly perturb- morphological features, and dependency labels. ing game score values without invalidating overall 5.5 Grammatical Error Correction (GEC) game outcome is one DA strategy explored in game summary generation (Hayashi et al., 2019). Lack of parallel data is typically a barrier for GEC. Two popular recent benchmarks are E2E-NLG Various works have thus looked at DA methods (Dušek et al., 2018) and WebNLG (Gardent et al., for GEC. We discuss some here, and more can be 2017). Both involve generation from structured found in Table 2 in Appendix C. inputs - meaning representation (MR) sequences There is work that makes use of additional re- and triple sequences, respectively. Montella et al. sources. Boyd (2018) use German edits from (2020) show performance gains on WebNLG by Wikipedia revision history and use those relating DA using Wikipedia sentences as targets and to GEC as augmented training data. Zhang et al. parsed OpenIE triples as inputs. Tandon et al. (2019b) explore multi-task transfer, or the use of (2018) propose DA for E2E-NLG based on per- annotated data from other tasks. muting the input MR sequence. Kedzie and McK- There is also work that adds synthetic errors to eown (2019) inject Gaussian noise into a trained noise the text. Wang et al. (2019a) investigate two decoder’s hidden states and sample diverse aug- approaches: token-level perturbations and training mented examples from it. This sample-augment- error generation models with a filtering strategy retrain loop helps performance on E2E-NLG. to keep generations with sufficient errors. Grund- kiewicz et al. (2019) use confusion sets generated 5.8 Open-Ended & Conditional Generation by a spellchecker for noising. Choe et al. (2019) learn error patterns from small annotated samples There has been limited work on DA for open-ended along with POS-specific noising. and conditional text generation. Feng et al. (2020) There have also been approaches to improve the experiment with a suite of DA methods for finetun- diversity of generated errors. Wan et al. (2020) ing GPT-2 on a low-resource domain in attempts investigate noising through editing the latent repre- to improve the quality of generated continuations, sentations of grammatical sentences, and Xie et al. which they call G ENAUG. They find that WN- (2018) use a neural sequence transduction model H YPERS (WordNet hypernym replacement of key- and beam search noising procedures. words) and S YNTHETIC N OISE (randomly perturb- ing non-terminal characters in words) are useful, 5.6 Neural Machine Translation (NMT) and the quality of generated text improves to a peak There are many works which have investigated DA at ≈ 3x the original amount of training data. for NMT. We highlighted some in §3 and §4.1, 5.9 Dialogue e.g. (Sennrich et al., 2016; Fadaee et al., 2017; Xia et al., 2019). We discuss some further ones here, Most DA approaches for dialogue focus on task- and more can be found in Table 3 in Appendix C. oriented dialogue. We outline some below, and Wang et al. (2018a) propose S WITCH O UT, a more can be found in Table 4 in Appendix C. DA method that randomly replaces words in both Quan and Xiong (2019) present sentence and source and target sentences with other random word-level DA approaches for end-to-end task- words from their corresponding vocabularies. Gao oriented dialogue. Louvan and Magnini (2020) et al. (2019) introduce S OFT C ONTEXTUAL DA propose LIGHTWEIGHT AUGMENTATION, a set of that softly augments randomly chosen words in a word-span and sentence-level DA methods for low- sentence using a contextual mixture of multiple resource slot filling and intent classification. 974
Hou et al. (2018) present a seq2seq DA frame- augmentation) can predict DA performance, but it work to augment dialogue utterances for dialogue is unclear how these results might translate to NLP. language understanding (Young et al., 2013), in- Minimal benefit for pretrained models on in- cluding a diversity rank to produce diverse utter- domain data: With the popularization of large ances. Zhang et al. (2019c) propose MADA to pretrained language models, it has recently come to generate diverse responses using the property that light that a couple of previously effective DA tech- several valid responses exist for a dialogue context. niques for certain text classification tasks in English There is also DA work for spoken dialogue. Hou (Wei and Zou, 2019; Sennrich et al., 2016) provide et al. (2018), Kim et al. (2019), Zhao et al. (2019), little benefit for models like BERT and RoBERTa, and Yoo et al. (2019) investigate DA methods for di- which already achieve high performance on in- alogue and spoken language understanding (SLU), domain text classification (Longpre et al., 2020). including generative latent variable models. One hypothesis for this could be that using simple 5.10 Multimodal Tasks DA techniques provides little benefit when finetun- ing large pretrained transformers on tasks for which DA techniques have also been proposed for multi- examples are well-represented in the pretraining modal tasks where aligned data for multiple modal- data, but DA methods could still be effective when ities is required. We look at ones that involve lan- finetuning on tasks for which examples are scarce guage or text. Some are discussed below, and more or out-of-domain compared with the training data. can be found in Table 5 in Appendix C. Further work could study under which scenarios Beginning with speech, Wang et al. (2020) pro- data augmentation for large pretrained models is pose a DA method to improve the robustness of likely to be effective. downstream dialogue models to speech recognition errors. Wiesner et al. (2018) and Renduchintala Multimodal challenges: While there has been et al. (2018) propose DA methods for end-to-end increased work in multimodal DA, as discussed in automatic speech recognition (ASR). §5.10, effective DA methods for multiple modal- Looking at images or video, Xu et al. (2020) ities has been challenging. Many works focus on learn a cross-modality matching network to pro- augmenting a single modality or multiple ones sep- duce synthetic image-text pairs for multimodal clas- arately. For example, there is potential to further sifiers. Atliha and Šešok (2020) explore DA meth- explore simultaneous image and text augmentation ods such as synonym replacement and contextual- for image captioning, such as a combination of ized word embeddings augmentation using BERT C UT M IX (Yun et al., 2019) and caption editing. for image captioning. Kafle et al. (2017), Yokota Span-based tasks offer unique DA challenges and Nakayama (2018), and Tang et al. (2020) pro- as there are typically many correlated classification pose methods for visual QA including question decisions. For example, random token replacement generation and adversarial examples. may be a locally acceptable DA method but possi- bly disrupt coreference chains for latter sentences. 6 Challenges & Future Directions DA techniques here must take into account depen- Looking forward, data augmentation faces substan- dencies between different locations in the text. tial challenges, specifically for NLP, and with these Working in specialized domains such as those challenges, new opportunities for future work arise. with domain-specific vocabulary and jargon (e.g. medicine) can present challenges. Many pretrained Dissonance between empirical novelties and models and external knowledge (e.g. WordNet) theoretical narrative: There appears to be a con- cannot be effectively used. Studies have shown spicuous lack of research on why DA works. Most that DA becomes less beneficial when applied to studies might show empirically that a DA technique out-of-domain data, likely because the distribution works and provide some intuition, but it is currently of augmented data can substantially differ from the challenging to measure the goodness of a technique original data (Zhang et al., 2019a; Herzig et al., without resorting to a full-scale experiment. A re- 2020; Campagna et al., 2020; Zhong et al., 2020). cent work in vision (Gontijo-Lopes et al., 2020) has proposed that affinity (the distributional shift Working with low-resource languages may caused by DA) and diversity (the complexity of the present similar difficulties as specialized domains. 975
Further, DA techniques successful in the high- in NLP could be fruitful, though another challenge resource scenario may not be effective for low- will be determining when such a module will be resource languages that are of a different language helpful, which—compared with CV, where the in- family or very distinctive in linguistic and typolog- variances being imposed are well-accepted—can ical terms. For example, those which are language vary substantially across NLP tasks. isolates or lack high-resource cognates. Lack of unification is a challenge for the cur- More vision-inspired techniques: Although rent literature on data augmentation for NLP, and many NLP DA methods have been inspired by anal- popular methods are often presented in an aux- ogous approaches in CV, there is potential for draw- iliary fashion. Whereas there are well-accepted ing further connections. Many CV DA techniques frameworks for DA for CV (e.g. default augmen- motivated by real-world invariances (e.g. many tation libraries in PyTorch, RandAugment (Cubuk angles of looking at the same object) may have et al., 2019)), there are no such "generalized" DA similar NLP interpretations. For instance, grayscal- techniques for NLP. Further, we believe that DA ing could translate to toning down aspects of the research would benefit from the establishment of text (e.g. plural to singular, "awesome" → "good"). standard and unified benchmark tasks and datasets Morphing a dependency tree could be analogous to compare different augmentation methods. to rotating an image, and paraphrasing techniques Good data augmentation practices would help may be analogous to changing perspective. For ex- make DA work more accessible and reproducible ample, negative data augmentation (NDA) (Sinha to the NLP and ML communities. On top of et al., 2021) involves creating out-of-distribution unified benchmark tasks, datasets, and frame- samples. It has so far been exclusively explored for works/libraries mentioned above, other good prac- CV, but could be investigated for text. tices include making code and augmented datasets Self-supervised learning: More recently, DA publicly available, reporting variation among re- has been increasingly used as a key component sults (e.g. standard deviation across random seeds), of self-supervised learning, particularly in vision and more standardized evaluation procedures. Fur- (Chen et al., 2020e). In NLP, BART (Lewis et al., ther, transparent hyperparameter analysis, explic- 2020) showed that predicting deleted tokens as a itly stating failure cases of proposed techniques, pretraining task can achieve similar performance as and discussion of the intuition and theory behind the masked LM, and E LECTRA (Clark et al., 2020) them would further improve the transparency and found that pretraining by predicting corrupted to- interpretability of DA techniques. kens outperforms BERT given the same model size, data, and compute. We expect future work will 7 Conclusion continue exploring how to effectively manipulate In this paper, we presented a comprehensive and text for both pretraining and downstream tasks. structured survey of data augmentation for nat- Offline versus online data augmentation: In ural language processing (NLP). We provided a CV, standard techniques such as cropping, color background about data augmentation and how it jittering, and rotations are typically done stochasti- works, discussed major methodologically represen- cally, allowing for DA to be incorporated elegantly tative data augmentation techniques for NLP, and into the training pipeline. In NLP, however, it is un- touched upon data augmentation techniques for clear how to include a lightweight code module to popular NLP applications and tasks. Finally, we apply DA stochastically. This is because DA tech- outlined current challenges and directions for fu- niques for NLP often leverage external resources ture research, and showed that there is much room (e.g. a word dictionary for token substitution or a for further exploration. Overall, we hope our paper translation model for backtranslation) that are not can serve as a guide for NLP researchers to decide easily transferable across model training pipelines. on which data augmentation techniques to use, and Thus, a common practice for DA in NLP is simply inspire additional interest and work in this area. to generate augmented data offline and store it as Please see the corresponding GitHub repository at additional data to be loaded during training.4 Fu- https://github.com/styfeng/DataAug4NLP. ture work on a lightweight module for online DA 4 See Appendix D. 976
References synthetic instance generation. Knowledge-Based Systems. Diego Alves, Askars Salimbajevs, and Mārcis Pinnis. 2020. Data augmentation for pipeline-based speech Nitesh V. Chawla, Kevin W. Bowyer, Lawrence O. Hall, translation. In 9th International Conference on Hu- and W. Philip Kegelmeyer. 2002. SMOTE: Syn- man Language Technologies - the Baltic Perspective thetic minority over-sampling technique. Journal of (Baltic HLT 2020), Kaunas, Lithuania. Artificial Intelligence Research, 16:321–357. Ateret Anaby-Tavor, Boaz Carmeli, Esther Goldbraich, Guanhua Chen, Yun Chen, Yong Wang, and Victor O.K. Amir Kantor, George Kour, Segev Shlomov, Naama Li. 2020a. Lexical-constraint-aware neural machine Tepper, and Naama Zwerdling. 2020. Do not have translation via data augmentation. In Proceedings of enough data? Deep learning to the rescue! In Pro- the Twenty-Ninth International Joint Conference on ceedings of AAAI, pages 7383–7390. Artificial Intelligence, IJCAI-20, pages 3587–3593. International Joint Conferences on Artificial Intelli- Jacob Andreas. 2020. Good-enough compositional gence Organization. Main track. data augmentation. In Proceedings of the 58th An- nual Meeting of the Association for Computational Hannah Chen, Yangfeng Ji, and David Evans. 2020b. Linguistics, pages 7556–7566, Online. Association Finding friends and flipping frenemies: Automatic for Computational Linguistics. paraphrase dataset augmentation using graph theory. In Findings of the Association for Computational Akari Asai and Hannaneh Hajishirzi. 2020. Logic- Linguistics: EMNLP 2020, pages 4741–4751, On- guided data augmentation and regularization for con- line. Association for Computational Linguistics. sistent question answering. In Proceedings of the 58th Annual Meeting of the Association for Compu- Jiaao Chen, Zichao Yang, and Diyi Yang. 2020c. Mix- tational Linguistics, pages 5642–5650, Online. As- Text: Linguistically-informed interpolation of hid- sociation for Computational Linguistics. den space for semi-supervised text classification. In Proceedings of the 58th Annual Meeting of the Asso- Viktar Atliha and Dmitrij Šešok. 2020. Text augmen- ciation for Computational Linguistics, pages 2147– tation using BERT for image captioning. Applied 2157, Online. Association for Computational Lin- Sciences, 10:5978. guistics. Christopher Beckham, Sina Honari, Vikas Verma, Shuxiao Chen, Edgar Dobriban, and Jane Lee. 2020d. Alex M. Lamb, Farnoosh Ghadiri, R Devon Hjelm, A group-theoretic framework for data augmentation. Yoshua Bengio, and Chris Pal. 2019. On adversarial Advances in Neural Information Processing Systems, mixup resynthesis. In Advances in Neural Informa- 33. tion Processing Systems, pages 4346–4357. Chris M. Bishop. 1995. Training with noise is equiv- Ting Chen, Simon Kornblith, Mohammad Norouzi, alent to Tikhonov regularization. Neural Computa- and Geoffrey Hinton. 2020e. A simple framework tion, 7(1):108–116. for contrastive learning of visual representations. In Proceedings of the 37th International Conference Adriane Boyd. 2018. Using Wikipedia edits in low on Machine Learning, volume 119 of Proceedings resource grammatical error correction. In Proceed- of Machine Learning Research, pages 1597–1607. ings of the 2018 EMNLP Workshop W-NUT: The PMLR. 4th Workshop on Noisy User-generated Text, pages 79–84, Brussels, Belgium. Association for Compu- Yong Cheng, Lu Jiang, Wolfgang Macherey, and Ja- tational Linguistics. cob Eisenstein. 2020. AdvAug: Robust adversar- ial augmentation for neural machine translation. In Bram Bulte and Arda Tezcan. 2019. Neural fuzzy re- Proceedings of the 58th Annual Meeting of the Asso- pair: Integrating fuzzy matches into neural machine ciation for Computational Linguistics, pages 5961– translation. In Proceedings of the 57th Annual Meet- 5970, Online. Association for Computational Lin- ing of the Association for Computational Linguistics, guistics. pages 1800–1809, Florence, Italy. Association for Computational Linguistics. Mara Chinea-Ríos, Álvaro Peris, and Francisco Casacuberta. 2017. Adapting neural machine trans- Giovanni Campagna, Agata Foryciarz, Mehrad Morad- lation with parallel synthetic data. In Proceedings shahi, and Monica Lam. 2020. Zero-shot transfer of the Second Conference on Machine Translation, learning with synthesized data for multi-domain dia- pages 138–147, Copenhagen, Denmark. Association logue state tracking. In Proceedings of the 58th An- for Computational Linguistics. nual Meeting of the Association for Computational Linguistics, pages 122–132, Online. Association for Yo Joong Choe, Jiyeon Ham, Kyubyong Park, and Computational Linguistics. Yeoil Yoon. 2019. A neural grammatical error cor- rection system built on better pre-training and se- F. Charte, Antonio Rivera Rivas, María José Del Je- quential transfer learning. In Proceedings of the sus, and Francisco Herrera. 2015. Mlsmote: Ap- Fourteenth Workshop on Innovative Use of NLP for proaching imbalanced multilabel learning through Building Educational Applications, pages 213–227, 977
Florence, Italy. Association for Computational Lin- on Natural Language Generation, pages 322–328, guistics. Tilburg University, The Netherlands. Association for Computational Linguistics. Kevin Clark, Minh-Thang Luong, Quoc V. Le, and Christopher D. Manning. 2020. ELECTRA: Pre- Alexander R. Fabbri, Simeng Han, Haoyuan Li, training text encoders as discriminators rather than Haoran Li, Marjan Ghazvininejad, Shafiq Joty, generators. Proceedings of ICLR. Dragomir Radev, and Yashar Mehdad. 2020. Im- proving zero and few-shot abstractive summariza- Ekin D. Cubuk, Barret Zoph, Dandelion Mane, Vijay tion with intermediate fine-tuning and data augmen- Vasudevan, and Quoc V. Le. 2018. Autoaugment: tation. arXiv preprint. Learning augmentation policies from data. Proceed- ings of CVPR. Marzieh Fadaee, Arianna Bisazza, and Christof Monz. 2017. Data augmentation for low-resource neural Ekin D. Cubuk, Barret Zoph, Jonathon Shlens, and machine translation. In Proceedings of the 55th An- Quoc V. Le. 2019. Randaugment: Practical data nual Meeting of the Association for Computational augmentation with no separate search. CoRR, Linguistics (Volume 2: Short Papers), pages 567– abs/1909.13719. 573, Vancouver, Canada. Association for Computa- tional Linguistics. Xiang Dai and Heike Adel. 2020. An analysis of simple data augmentation for named entity recogni- Alex Falcon, Oswald Lanz, and Giuseppe Serra. 2020. tion. In Proceedings of the 28th International Con- Data augmentation techniques for the video question ference on Computational Linguistics, pages 3861– answering task. arXiv preprint. 3867, Barcelona, Spain (Online). International Com- Mojtaba Faramarzi, Mohammad Amini, Akilesh Badri- mittee on Computational Linguistics. naaraayanan, Vikas Verma, and Sarath Chandar. Hao T. Dang. 2005. Overview of DUC 2005. 2020. Patchup: A regularization technique for con- volutional neural networks. arXiv preprint. Tri Dao, Albert Gu, Alexander J. Ratner, Virginia Mariano Felice. 2016. Artificial error generation for Smith, Christopher De Sa, and Christopher Ré. 2019. translation-based grammatical error correction. Uni- A kernel theory of modern data augmentation. Pro- versity of Cambridge Technical Report. ceedings of Machine Learning Research, 97:1528. Steven Y. Feng, Varun Gangal, Dongyeop Kang, Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Teruko Mitamura, and Eduard Hovy. 2020. GenAug: Kristina Toutanova. 2019. BERT: Pre-training of Data augmentation for finetuning text generators. In deep bidirectional transformers for language under- Proceedings of Deep Learning Inside Out (DeeLIO): standing. In Proceedings of the 2019 Conference of The First Workshop on Knowledge Extraction and the North American Chapter of the Association for Integration for Deep Learning Architectures, pages Computational Linguistics: Human Language Tech- 29–42, Online. Association for Computational Lin- nologies, Volume 1 (Long and Short Papers), pages guistics. 4171–4186. Steven Y. Feng, Aaron W. Li, and Jesse Hoey. 2019. Terrance DeVries and Graham W Taylor. 2017. Im- Keep calm and switch on! Preserving sentiment and proved regularization of convolutional neural net- fluency in semantic text exchange. In Proceedings works with cutout. arXiv preprint. of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th Interna- Bosheng Ding, Linlin Liu, Lidong Bing, Canasai Kru- tional Joint Conference on Natural Language Pro- engkrai, Thien Hai Nguyen, Shafiq Joty, Luo Si, and cessing (EMNLP-IJCNLP), pages 2701–2711, Hong Chunyan Miao. 2020. DAGA: Data augmentation Kong, China. Association for Computational Lin- with a generation approach forLow-resource tagging guistics. tasks. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Process- Alberto Fernández, Salvador Garcia, Francisco Her- ing (EMNLP), pages 6045–6057, Online. Associa- rera, and Nitesh V. Chawla. 2018. SMOTE for learn- tion for Computational Linguistics. ing from imbalanced data: progress and challenges, marking the 15-year anniversary. Journal of Artifi- Tommaso Dreossi, Shromona Ghosh, Xiangyu Yue, cial Intelligence Research, 61:863–905. Kurt Keutzer, Alberto L. Sangiovanni-Vincentelli, and Sanjit A. Seshia. 2018. Counterexample-guided Jennifer Foster and Oistein Andersen. 2009. Gen- data augmentation. In Proceedings of IJCAI. ERRate: Generating errors for use in grammatical error detection. In Proceedings of the Fourth Work- Sufeng Duan, Hai Zhao, Dongdong Zhang, and Rui shop on Innovative Use of NLP for Building Edu- Wang. 2020. Syntax-aware data augmentation for cational Applications, pages 82–90, Boulder, Col- neural machine translation. arXiv preprint. orado. Association for Computational Linguistics. Ondřej Dušek, Jekaterina Novikova, and Verena Rieser. Varun Gangal, Steven Y. Feng, Eduard Hovy, and 2018. Findings of the E2E NLG challenge. In Teruko Mitamura. 2021. Nareor: The narrative re- Proceedings of the 11th International Conference ordering problem. arXiv preprint. 978
Fei Gao, Jinhua Zhu, Lijun Wu, Yingce Xia, Tao Kristina Gulordava, Piotr Bojanowski, Edouard Grave, Qin, Xueqi Cheng, Wengang Zhou, and Tie-Yan Liu. Tal Linzen, and Marco Baroni. 2018. Colorless 2019. Soft contextual data augmentation for neural green recurrent networks dream hierarchically. In machine translation. In Proceedings of the 57th An- Proceedings of the 2018 Conference of the North nual Meeting of the Association for Computational American Chapter of the Association for Computa- Linguistics, pages 5539–5544, Florence, Italy. Asso- tional Linguistics: Human Language Technologies, ciation for Computational Linguistics. Volume 1 (Long Papers), pages 1195–1205, New Orleans, Louisiana. Association for Computational Silin Gao, Yichi Zhang, Zhijian Ou, and Zhou Yu. Linguistics. 2020. Paraphrase augmented task-oriented dialog generation. In Proceedings of the 58th Annual Meet- Demi Guo, Yoon Kim, and Alexander Rush. 2020. ing of the Association for Computational Linguistics, Sequence-level mixed sample data augmentation. In pages 639–649, Online. Association for Computa- Proceedings of the 2020 Conference on Empirical tional Linguistics. Methods in Natural Language Processing (EMNLP), pages 5547–5552, Online. Association for Computa- tional Linguistics. Claire Gardent, Anastasia Shimorina, Shashi Narayan, and Laura Perez-Beltrachini. 2017. The WebNLG Hongyu Guo. 2020. Nonlinear mixup: Out-of- challenge: Generating text from RDF data. In Pro- manifold data augmentation for text classification. ceedings of the 10th International Conference on In Proceedings of AAAI, pages 4044–4051. Natural Language Generation, pages 124–133, San- tiago de Compostela, Spain. Association for Compu- Rowan Hall Maudslay, Hila Gonen, Ryan Cotterell, tational Linguistics. and Simone Teufel. 2019. It’s all in the name: Mit- igating gender bias with name-based counterfactual Golnaz Ghiasi, Yin Cui, Aravind Srinivas, Rui Qian, data substitution. In Proceedings of the 2019 Con- Tsung-Yi Lin, Ekin D. Cubuk, Quoc V. Le, and ference on Empirical Methods in Natural Language Barret Zoph. 2020. Simple copy-paste is a strong Processing and the 9th International Joint Confer- data augmentation method for instance segmenta- ence on Natural Language Processing (EMNLP- tion. arXiv preprint. IJCNLP), pages 5267–5275, Hong Kong, China. As- sociation for Computational Linguistics. Max Glockner, Vered Shwartz, and Yoav Goldberg. 2018. Breaking NLI systems with sentences that re- Bharath Hariharan and Ross Girshick. 2017. Low- quire simple lexical inferences. In Proceedings of shot visual recognition by shrinking and hallucinat- the 56th Annual Meeting of the Association for Com- ing features. In Proceedings of the IEEE Interna- putational Linguistics (Volume 2: Short Papers), tional Conference on Computer Vision, pages 3018– pages 650–655, Melbourne, Australia. Association 3027. for Computational Linguistics. Hany Hassan, Mostafa Elaraby, and Ahmed Tawfik. 2017. Synthetic data for neural machine translation Raphael Gontijo-Lopes, Sylvia J. Smullin, Ekin D. of spoken-dialects. arXiv preprint. Cubuk, and Ethan Dyer. 2020. Tradeoffs in data augmentation: An empirical study. Proceedings of Hiroaki Hayashi, Yusuke Oda, Alexandra Birch, Ioan- ICLR. nis Konstas, Andrew Finch, Minh-Thang Luong, Graham Neubig, and Katsuhito Sudoh. 2019. Find- Miguel Graça, Yunsu Kim, Julian Schamper, Shahram ings of the third workshop on neural generation and Khadivi, and Hermann Ney. 2019. Generalizing translation. In Proceedings of the 3rd Workshop back-translation in neural machine translation. In on Neural Generation and Translation, pages 1–14, Proceedings of the Fourth Conference on Machine Hong Kong. Association for Computational Linguis- Translation (Volume 1: Research Papers), pages 45– tics. 52, Florence, Italy. Association for Computational Linguistics. Karl Moritz Hermann, Tomas Kocisky, Edward Grefen- stette, Lasse Espeholt, Will Kay, Mustafa Suleyman, Milan Gritta, Gerasimos Lampouras, and Ignacio Ia- and Phil Blunsom. 2015. Teaching machines to read cobacci. 2021. Conversation graph: Data augmen- and comprehend. In NeurIPS, pages 1693–1701. tation, training, and evaluation for non-deterministic Alex Hernández-García and Peter König. 2020. Data dialogue management. Transactions of the Associa- augmentation instead of explicit regularization. tion for Computational Linguistics, 9:36–52. arXiv preprint. Roman Grundkiewicz, Marcin Junczys-Dowmunt, and Jonathan Herzig, Pawel Krzysztof Nowak, Thomas Kenneth Heafield. 2019. Neural grammatical error Müller, Francesco Piccinno, and Julian Eisenschlos. correction systems with unsupervised pre-training 2020. TaPas: Weakly supervised table parsing via on synthetic data. In Proceedings of the Fourteenth pre-training. In Proceedings of the 58th Annual Workshop on Innovative Use of NLP for Building Meeting of the Association for Computational Lin- Educational Applications, pages 252–263, Florence, guistics, pages 4320–4333, Online. Association for Italy. Association for Computational Linguistics. Computational Linguistics. 979
Yutai Hou, Yijia Liu, Wanxiang Che, and Ting Liu. Dongyeop Kang, Tushar Khot, Ashish Sabharwal, and 2018. Sequence-to-sequence data augmentation for Eduard Hovy. 2018. AdvEntuRe: Adversarial train- dialogue language understanding. In Proceedings of ing for textual entailment with knowledge-guided ex- the 27th International Conference on Computational amples. In Proceedings of the 56th Annual Meet- Linguistics, pages 1234–1245, Santa Fe, New Mex- ing of the Association for Computational Linguistics ico, USA. Association for Computational Linguis- (Volume 1: Long Papers), pages 2418–2428, Mel- tics. bourne, Australia. Association for Computational Linguistics. Jian Huang, Ya Li, Jianhua Tao, Zheng Lian, Mingyue Niu, and Minghao Yang. 2018. Multimodal continu- Min-Hyung Kang. 2019. Valar nmt : Vastly lack- ous emotion recognition with data augmentation us- ing resources neural machine translation. Stanford ing recurrent neural networks. In Proceedings of the CS224N. 2018 on Audio/Visual Emotion Challenge and Work- shop, AVEC’18, page 57–64, New York, NY, USA. Omid Kashefi and Rebecca Hwa. 2020. Quantifying Association for Computing Machinery. the evaluation of heuristic methods for textual data augmentation. In Proceedings of the Sixth Work- Brian Kenji Iwana and Seiichi Uchida. 2020. An em- shop on Noisy User-generated Text (W-NUT 2020), pirical survey of data augmentation for time series pages 200–208, Online. Association for Computa- classification with neural networks. arXiv preprint. tional Linguistics. Mohit Iyyer, John Wieting, Kevin Gimpel, and Luke Divyansh Kaushik, Eduard Hovy, and Zachary Lipton. Zettlemoyer. 2018. Adversarial example generation 2020. Learning the difference that makes a differ- with syntactically controlled paraphrase networks. ence with counterfactually-augmented data. In Inter- In Proceedings of the 2018 Conference of the North national Conference on Learning Representations. American Chapter of the Association for Computa- tional Linguistics: Human Language Technologies, Divyansh Kaushik, Amrith Setlur, Eduard H. Hovy, Volume 1 (Long Papers), pages 1875–1885, New and Zachary Chase Lipton. 2021. Explaining the ef- Orleans, Louisiana. Association for Computational ficacy of counterfactually augmented data. In Inter- Linguistics. national Conference on Learning Representations. Robin Jia and Percy Liang. 2016. Data recombination Chris Kedzie and Kathleen McKeown. 2019. A good for neural semantic parsing. In Proceedings of the sample is hard to find: Noise injection sampling and 54th Annual Meeting of the Association for Compu- self-training for neural language generation models. tational Linguistics (Volume 1: Long Papers), pages In Proceedings of the 12th International Conference 12–22, Berlin, Germany. Association for Computa- on Natural Language Generation, pages 584–593, tional Linguistics. Tokyo, Japan. Association for Computational Lin- Robin Jia, Aditi Raghunathan, Kerem Göksel, and guistics. Percy Liang. 2019. Certified robustness to adver- sarial word substitutions. In Proceedings of the Hwa-Yeon Kim, Yoon-Hyung Roh, and Young-Kil 2019 Conference on Empirical Methods in Natu- Kim. 2019. Data augmentation by data noising ral Language Processing and the 9th International for open-vocabulary slots in spoken language under- Joint Conference on Natural Language Processing standing. In Proceedings of the 2019 Conference of (EMNLP-IJCNLP), pages 4120–4133. the North American Chapter of the Association for Computational Linguistics: Student Research Work- Amit Jindal, Arijit Ghosh Chowdhury, Aniket Didolkar, shop, pages 97–102, Minneapolis, Minnesota. Asso- Di Jin, Ramit Sawhney, and Rajiv Ratn Shah. 2020a. ciation for Computational Linguistics. Augmenting NLP models using latent feature in- terpolations. In Proceedings of the 28th Inter- Alex Kimn. 2020. A syntactic rule-based framework national Conference on Computational Linguistics, for parallel data synthesis in japanese gec. Mas- pages 6931–6936, Barcelona, Spain (Online). Inter- sachusetts Institute of Technology. national Committee on Computational Linguistics. Sosuke Kobayashi. 2018. Contextual augmentation: Amit Jindal, Narayanan Elavathur Ranganatha, Aniket Data augmentation by words with paradigmatic re- Didolkar, Arijit Ghosh Chowdhury, Di Jin, Ramit lations. In Proceedings of the 2018 Conference of Sawhney, and Rajiv Ratn Shah. 2020b. Speechmix the North American Chapter of the Association for - augmenting deep sound recognition using hidden Computational Linguistics: Human Language Tech- space interpolations. In INTERSPEECH, pages 861– nologies, Volume 2 (Short Papers), pages 452–457, 865. New Orleans, Louisiana. Association for Computa- tional Linguistics. Kushal Kafle, Mohammed Yousefhussien, and Christo- pher Kanan. 2017. Data augmentation for visual Ashutosh Kumar, Kabir Ahuja, Raghuram Vadapalli, question answering. In Proceedings of the 10th In- and Partha Talukdar. 2020. Syntax-guided con- ternational Conference on Natural Language Gen- trolled generation of paraphrases. Transactions eration, pages 198–202, Santiago de Compostela, of the Association for Computational Linguistics, Spain. Association for Computational Linguistics. 8:329–345. 980
You can also read