PENS: A Dataset and Generic Framework for Personalized News Headline Generation
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
PENS: A Dataset and Generic Framework for Personalized News Headline Generation Xiang Ao♠♦∗, Xiting Wang♥ , Ling Luo♠♦ , Ying Qiao♣ , Qing He♠♦ , Xing Xie♥† ♠ Key Lab of Intelligent Information Processing of Chinese Academy of Sciences (CAS), Institute of Computing Technology, CAS, Beijing 100190, China ♥ Microsoft Research Asia ♣ Microsoft ♦ University of Chinese Academy of Sciences, Beijing 100049, China {aoxiang,luoling18s,heqing}@ict.ac.cn {xitwan,yiqia,xing.xie}@microsoft.com Abstract etc (LaRocque, 2003) propels a new research di- rection that how to decorate the headline as an irre- In this paper, we formulate the personalized sistible invitation to users for reading through the news headline generation problem whose goal is to output a user-specific title based on both article (Xu et al., 2019) since more readings may ac- a user’s reading interests and a candidate news quaint more revenue of these platforms. To this end, body to be exposed to her. To build up a specified stylized headline generation techniques benchmark for this problem, we publicize a were proposed, such as question headline (Zhang large-scale dataset named PENS (PErsonal- et al., 2018), sensational headline (Xu et al., 2019) ized News headlineS). The training set is col- generation, and so on (Shu et al., 2018; Gu et al., lected from user impressions logs of Microsoft 2020). However, the over-decorate headlines might News, and the test set is manually created bring negative effects as click-baits begin to be- by hundreds of native speakers to enable a fair testbed for evaluating models in an offline come notorious in ubiquitous online services1 . mode. We propose a generic framework as Hence, the question is now changing to how a preparatory solution to our problem. At its to construct a title that catches on reader curios- heart, user preference is learned by leveraging ity without entering into click-bait territory. In- the user behavioral data, and three kinds of spired by the tremendous success of personalized user preference injections are proposed to per- news recommendation (An et al., 2019; Wang et al., sonalize a text generator and establish person- 2018; Li et al., 2010; Zheng et al., 2018) where the alized headlines. We investigate our dataset by implementing several state-of-the-art user ultimate goal is to learn users’ reading interests modeling methods in our framework to demon- and deliver the right news to them, a plausible solu- strate a benchmark score for the proposed tion to this question could be producing headlines dataset. The dataset is available at https: satisfying the personalized interests of readers. //msnews.github.io/pens.html. It thus motivates the study of the personalized news headline generation whose goal is to output a 1 Introduction user-specific title based on both a user’s reading in- News headline generation (Dorr et al., 2003; Lopy- terests and a candidate news body to be exposed to rev, 2015; Alfonseca et al., 2013; Tan et al., 2017; her. Analogous to personalized news recommenda- See et al., 2017; Zhang et al., 2018; Xu et al., 2019; tions, user preference can be learned by leveraging Murao et al., 2019; Gavrilov et al., 2019; Gu et al., the behavioral data of readers on content vendors, 2020; Song et al., 2020), conventionally considered and the representation could personalize text gen- as a paradigm of challenging text summarization erators and establish distinct headlines, even with task, has been extensively explored for decades. the same news body, for different readers. Their intuitive intention is to empower the model However, it might be difficult to evaluate the ap- to output a condensed generalization, e.g., one sen- proaches of personalized headline generation due tence, of a news article. to the lack of large-scale available datasets. First, The recent year escalation of online content there are few available benchmarks that simultane- vendors such as Google News, TopBuzz, and ously contain user behavior and news content to ∗ train models. For example, most available news rec- This work was done when Xiang was visiting MSRA supported by the MSRA Young Visiting Researcher Program. 1 https://www.vizion.com/blog/ † Corresponding author. do-clickbait-titles-still-work/ 82 Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, pages 82–92 August 1–6, 2021. ©2021 Association for Computational Linguistics
ommendation datasets may predominately contain Click Probability user-side interaction data, e.g., exposure impres- Dot rv ru Click Predictor sions and click behaviors, but the textual features usually have already been overly pre-processed (Li User Encoder et al., 2010; Zheng et al., 2018). As a result, ad- News News … News vanced NLP techniques that extract useful features News Encoder Encoder Encoder Encoder from textual data are limited. News headline gen- eration datasets, on the other hand, usually consist … of news bodies as well as their headlines, which all Candidate News Clicked News come from the news-side (Tan et al., 2017; Zhang Figure 1: Personalized news recommendation frame- et al., 2018) rather than the user-side. Though the work. MIND dataset (Wu et al., 2020), which was pre- sented by Microsoft, simultaneously contains the user-side behavioral data and the news-side origi- 2017) decoder. We implement six state-of-the-arts nal textual data, it was constructed for personalized personalized news recommendation approaches to news recommendations rather than our problem. model user preferences and provide a horizontal The more challenging issue for evaluating person- standard for the PENS dataset. The experimen- alized headline generation approaches is the severe tal results show effective personalization modeling cost during the test phase. It could be intractable and comprehensive injection of user interests can and infeasible to do an A/B test for every model in underpin an improvement in the quality of person- online environments. An efficient and fair testbed alized news headline generation. We expect PENS to evaluate the models in an offline mode is in can serve as a benchmark for personalized headline urgent demand to make the effectiveness and repro- generation and bolster the research in this area. ducibility of proposed models comparable. To this end, we publicize a dataset named 2 Problem Formulation and Discussion PENS (PErsonalized News headlineS) in this pa- In this section, we formulate the problem of person- per as a benchmark to testify the performance of alized news headline generation and differentiate it personalized news headline generation approaches. from personalized news recommendations. The training set of PENS is collected from the user impression logs of Microsoft News2 , in which 2.1 Problem Formulation 500, 000 impressions over 445, 765 users on more than one hundred thousand English news articles The problem of personalized news headline gen- are provided. In addition, we collected 103 English eration is formulated as follows. Given a user u native speakers’ click behaviors as well as their on an online content vendor, we denote his past more than 20, 000 manually-crafted personalized click history as [cu1 , cu2 , . . . , cuN ] where each c rep- headlines of news articles on the same news corpus resents the headline of user u’s clicked news and for testing. These manually-written headlines are each headline is composed of a sequence of words regarded as the gold standard of the user-preferred c = [wc1 , . . . , wcT ] with the maximum length of titles. Then, proposed methods can take prevailing T . Then, given the news body of a piece of news matching metrics, e.g., ROUGE, BLEU and etc., v = [wv1 , . . . , wvn ] to be exposed to user u, our to verify the performance. problem is to generate a personalized news head- line Hvu = [yvu1 , . . . , yvuT ] based on the clicked Moreover, we propose a generic framework to news [cu1 , cu2 , . . . , cuN ] and v. inject personalized interests into a proposed neural headline generator to enable a beacon for this area, 2.2 Difference to Personalized News considering there are few existing works that can Recommendation generate personalized news headlines. In more de- tail, we devise three kinds of incorporation methods Here we differentiate our problem from personal- to inject user interest representation into a proposed ized news recommendation whose general frame- neural headline generator with a transformer-based work is shown as Fig. 1. encoder and a pointer network-based (See et al., Recall that the aim of personalized news rec- ommendation is computing and matching between 2 https://microsoftnews.msn.com the candidate news and the user’s interests. Hence, 83
(a) Distribution of title length (b) Distribution news body length 0.14 3 PENS Dataset 0.4 0.12 0.10 0.3 In this section, we detail our PENS dataset. The 0.08 0.06 0.2 dataset was randomly sampled impression logs of 0.04 0.1 Microsoft News from June 14 to July 12, 2019. 0.02 0.00 0.0 15 2k Both user behaviors and news contents are involved, title length news body length (c) Distribution of entity number (d) Distribution of entity number and each user was de-linked from the production in news title in news body system when securely hashed into an anonymous 0.25 0.20 ID to reserve the data privacy issues. 0.20 0.15 0.15 0.10 3.1 News Corpus 0.10 0.05 0.05 The PENS dataset contains 113, 762 pieces of news 0.00 0.00 0 1 2 3 4 >4 30 entity number entity number articles whose topics are distributed into 15 cate- (f) Distribution of the length gories. The topical distribution is demonstrated in (e) Distribution of news topic of click history per user 0.35 Fig. 2 (c). Each news article in the PENS dataset 0.25 0.30 0.20 0.25 includes a news ID, a title, a body and a category 0.20 0.15 0.15 label. The average length of news title and news 0.10 0.05 0.10 body is 10.5 and 549.0, individually. Moreover, we 0.05 0.00 0.00 extract entities from each news title and body and 80 sports news F&D joy finance music lifestyle weather health video autos movies tv travel kids click number link them to the entities in WikiData3 . It could be Figure 2: The statistics of news corpus and training set taken as an auxiliary source to facilitate knowledge- of the PENS dataset. aware personalization modeling and headline gener- ation. The key statistical information of the PENS dataset is exhibited in Fig. 2 (a)–(e). 3.2 Training Set learning accurate news and user representations is critical for this problem. Under the neural frame- The training set of PENS consists of impression work, the news representation is usually modeled logs. An impression log records the news arti- by a news encoder that encodes news title, news cles displayed to a user as well as the click be- body or other attributes via various neural struc- haviors on these news articles when he/she visits tures (Okura et al., 2017; Wang et al., 2018; Wu the news website homepage at a specific time. We et al., 2019a; An et al., 2019; Wu et al., 2019a). follow the MIND dataset (Wu et al., 2020) that The user representation is generated by engrav- we add the news click histories of every individ- ing the high-level aspects over their clicked news ual user to his/her impression log to offer labeled sequences using sequential (Okura et al., 2017; samples for learning user preferences. Hence, the An et al., 2019) or attentive modules (Wu et al., format of each labeled sample in our training set is 2019b,a), in which every news is encoded by the [uID, tmp, clkNews, uclkNews, clkedHis], where news encoder in advance. Finally, the two repre- uID indicates the anonymous ID of a user, tmp sentations are matched by the click predictor, and denotes the timestamp of this impression record. the whole model is trained by the supervision of clkNews and uclkNews are the clicked news and click signals. un-clicked news in this impression, respectively. clkedHis represents the news articles previously Different from personalized news recommenda- clicked by this user. All the samples in clkNews, tions, our personalized news headline generation uclkNews and clkedHis are news IDs, and they all could be regarded as an NLP task than a user mod- sort by the user’s click time. The histogram of the eling and matching problem. Although it similarly number of news in the clicked history per user is needs to model preferences for the individual users shown in Fig. 2 (f). as what personalized news recommendations do, 3.3 Test Set the output of our problem is a natural language To provide an offline testbed, we invited 103 En- sequence that the target user might be interested glish native speakers (all are college students) man- in, i.e., user-preferred news title, rather than a click 3 probability score. https://www.wikidata.org/wiki/Wikidata:MainPage 84
Table 1: The statistics of the training and test set in PENS. “wd.” in the table means word. ×(1 − ptgen ) ×ptgen Final Distribution P(yt ) #impression #news #user Train 500,000 111,762 445,765 Context Vocabulary Test NA 60,000 103 Distribution Vector avg. click/user avg. wd./title avg. wd./body Pvocab (yt ) Train 74.8 10.5 549.0 ptgen Test 107.6 10.8 548.2 Attention Distribution ! F2 F1 Encoder ! ! Decoder Hidden States Hidden States hv 1 hv 2 hv n xt−1 xt " ! # ually create a test set by two stages. At the first stage, each person browses 1, 000 news headlines Transformer Encoder and marks at least 50 pieces he/she is interested in. User Embedding These exhibited news headlines were randomly se- lected from our news corpus and were arranged by w v1 w v2 ! wv n User Encoder their first exposure time. At the second stage, every- News News … News Encoder Encoder Encoder one is asked to write down their preferred headlines for another 200 news articles from our corpus, with- v cu1 cu2 … cuN out exhibiting them the original news titles. Note Candidate News Body Clicked News that these news articles are excluded from the first Figure 3: The generic framework of personalized news stage, and only news bodies were exhibited to these headline generation. Three kinds of user embedding annotators in this stage. These news articles are injections are devised. 1 : utilizing user embedding to evenly sampled, and we redundantly assign them initialize the decoder’s hidden state of the headline gen- to make sure each news is exhibited to four people erator. 2 : personalizing the attentive values on words on average. The quality of these manually-written in the news body by the user embedding. 3 : perturb- headlines was checked by professional editors from ing the choice between generation and copying via the the perspective of the factual aspect of the media user embedding. frame (Wagner and Gruszczynski, 2016). Low- quality headlines, e.g. containing wrong factual 4.1 Headline Generator information, inconsistent with the news body, too- The pin-point of our proposed headline generator short or overlong, etc., are removed. The rest are is a variant of transformer encoder and pointer net- regarded as the personalized reading focuses of work decoder. During the encoding, given the news these annotators on the articles and are taken as body of a candidate news v = [wv1 , . . . , wvn ], its gold-standard headlines in our dataset. The statis- word embeddings [ev1 , . . . , evn ] ∈ Rdw are first tics of the training and test sets of the PENS are fed to a two-layer positional encoder. The first shown in Table 1. layer aims to enhance the word structure within the whole news body sequence following Vaswani et al. 4 Our Framework (2017), and we add the positional encoding to each embedding vector with, In this section, we illustrate our generic framework PE (pos,2i) = sin(pos/100002i/dw ) (1) for resolving personalized news headline genera- tion, and its key issue is how to inject the user PE (pos,2i+1) = cos(pos/100002i/dw ) (2) preference into a news headline generator. We devise a headline generator with a transformer en- where pos is the word position and i is the dimen- coder and a pointer network decoder as our base sion. We also apply a sentence-layer positional en- model and propose three kinds of manners of in- coding to discover structural relations from higher jecting the user interests to generate personalized level. Suppose the Wpos ∈ RL×ds represents the headlines. The user interests can be derived fol- position embedding matrix of sentence level where lowing the approaches in news recommendations L is the sentence length and ds is the embedding community, and we omit its details due to the space size, the l-th row of Wpos represents the positional limitation. The architecture of our proposed frame- embedding of all the words in the l-th sentence. work is shown as Figure 3. Thus, each word embedding e0pos with positional 85
information can be represented as: where PvocabP(wt ) is zero when wt is out of vocab- e0pos = (epos + PE pos ) ⊕ Wpos [l]. (3) ulary while j:wj =wt at,j = 0 when the wt is not in the news body. ptgen is calculated based on the where ⊕ means concatenation. Furthermore, multi- context vector ct , decoder state st and the decoder head self-attention mechanism (Vaswani et al., input xt : 2017) is adopted to capture the word and sentence ptgen = Tθ (ct , st , xt ), (9) interactions by, where Tθ is a function template as Eq. (6). E 0 WiQ (E 0 WiK )> hi = softmax( √ )E 0 WiV (4) 4.2 Personalization by Injecting User dk Interests where dk = ds +d k w and i = 1, . . . , k given k So far, the imperative issue is to personalize the Q heads. Wi , Wi , Wi ∈ R(ds +dw )×dk . E 0 rep- K V headline generator by injecting the user’s prefer- resents the word sequence embeddings in candi- ence. Recall that we can obtain user embedding date news v. Thus, the encoder hidden states indicating user’s reading interests based on his/her h = h1 ⊕ h2 , . . . , hk can be derived. historical clicked news sequences, and we denote During the process of decoding, the decoded such representation as u. As the user embedding u hidden state st at time step t can be derived after is usually not aligned with the word embeddings, given the input xt , and an attention distribution at it remains challenges to incorporate the user in- over the encoder hidden states h is calculated as, terests to influence the headline generation with at = Fθ (h, st ) (5) personalized information. In our framework, based on our headline gener- ator, we propose three different manners to inject > Fθ (h, st ) = softmax(Vatt tanh(Wh h + Ws st + batt )) (6) user interests, considering different intuitions, and they are exhibited in Fig. 3. First, the most simple where Fθ represents a function template parame- and intuitive choice is to utilize the user embed- terized by θ to combine the linear transformation ding u to initialize the decoder hidden state of the of the encoder and the decoder states, i.e., h and headline generator. Second, under the empirical st . Next, the context vector ct , which can be seen assumption that users may attend on different para- as a fixed-size representation read from the news graphs and words in news articles corresponding body at time step t, is computed by a weighted to their individual preference, we inject u to affect sum of the encoder hidden states over the attention the attention distribution at in order to personal- distribution. Then the vocabulary distribution is ize the attentive values on the different words in produced by, the news body. That is, we modify Eq. (5) and Pvocab (wt ) = tanh(Vp [st ; ct ] + bv ), (7) derive at = Fθ (h, st , u). Lastly, we incorporate the personalized information to perturb the choice where Vp and bv are learnable parameters while between generating a word from vocabulary or Pvocab (wt ) represents the probability distribution copying a word from the news body, and derive over all the words in the vocabulary to predict the ptgen = Tθ (ct , st , xt , u). Compared with Eq. (9), u word at time step t. is taken as an auxiliary parameter, where Tθ is also Inspired by pointer-generator network (See et al., a function template as Eq. (6). 2017), which exhibits desirable performance on ei- ther dealing with out-of-vocabulary (OOV) words 4.3 Training or improving the reproducing factual details with In this subsection, we present the training process copy mechanism, we adopt a pointer ptgen at de- of our framework. The headline generation can coding step t as a soft switch to choose between be considered as a sequential decision-making pro- generating a word from the vocabulary with a prob- cess, hence we optimize a θ parametrized policy for ability of Pvocab (wt ) or copying a word from the the generator by maximizing the expected reward news body sampling from the attention distribu- of generated headline Y1:T : tion at . Thus, the probability distribution over the EY1:T ∼Gθ [R(Y1:T )]. (10) extended vocabulary is computed by, P (wt ) = ptgen Pvocab (wt ) + (1 − ptgen ) X at,j (8) For the generator, policy gradient methods are ap- j:wj =wt plied to maximize the objective function in Eq. (10), 86
whose gradient can be derived as, multi-head self-attention. (5) LSTUR (An et al., 2019) models long- and shor-term user represen- ∇θ J(θ) ' Eyt ∼Gθ (yt |Y1:t−1 ) (11) tations based on user ID embedding and sequen- [∇θ log Gθ (yt |Y1:t−1 ) · R(Y1:t−1 , yt )] tial encoding, individually. (6) NAML (Wu et al., where the reward R is estimated by the degree of 2019a) proposes multi-view learning in user repre- personalization, fluency and factualness as we aim sentation. to generate a user-specific and coherent headline to To the best of our knowledge, there are no cover the main theme of news articles and arouse exclusive methods for personalized news head- personalized reading curiosity. The implemented line generation. Hence we take several headline rewards in our framework contain: (1) The person- generation methods for comparison. (1) Pointer- alization of the generated headline is measured by Gen (See et al., 2017) proposes an explicit proba- the dot product between the user embedding and bilistic switch to choose between copying from the generated headline representation. Such a score source text and generating word from vocabu- might imply a matching degree of personalization. lary. (2) PG+RL-ROUGE (Xu et al., 2019) ex- (2) The fluency of a generated headline is assessed tends Pointer-Gen with as a reinforcement learning by a language model. We adopt a two-layer LSTM framework which generates sensational headlines pre-trained by maximizing the likelihood of news by considering ROUGE-L score as rewards. body and consider the probability estimation of a 5.2 Experiment Setup generated headline as the fluency reward. (3) We We perform the following preprocessings. For each measure the degree of factual consistency and the impression, we empirically keep at most 50 clicked coverage by calculating the mean of ROUGE (Lin, news to learn user preferences, and set the length of 2004)-1, -2 and -L F-scores between each sentence news headline and news body to 30 and 500, respec- in the news body and the generated headline, and tively. Word embeddings are 300-dimension and then take the average of the top 3 scores as the initialized by the Glove (Pennington et al., 2014) reward. We average all three rewards as the fi- while the size of position embeddings at sentence nal signal. As all the above reward functions only level is 100. The multi-head attention networks produce an end reward after the whole headline is have 8 heads. generated, we apply a Monte Carlo Tree search to First of all, we conduct news recommendation estimate the intermediate rewards. tasks to pretrain a user encoder with a learning rate 5 Experimental Evaluation of 10−4 on the first three weeks, i.e., from June 14 to July 4, 2019, on the training set, and test on In this section, we investigate our proposed PENS the rest. Notice that the parameters of the user en- dataset and conduct several comparisons to give coder are not updated thereafter. Meanwhile, the benchmark scores of personalized headline genera- headline generator is also pretrained with a learn- tion on this dataset. In the following part, we will ing rate of 0.001 by maximizing the likelihood of introduce the compared methods first, and then de- original headlines based on a random but fixed user tail the experimental setup, and finally present the embedding which can be considered as a global results and analysis. user without personalized information. Next, we train each individual model for 2 epochs follow- 5.1 Compared Methods ing Eq. 10, and Adam (Kingma and Ba, 2014) is We mainly compare two groups of approaches. The used for model optimization where we sample 16 first group consists of various user modeling meth- sequences for Monte Carlo search. ods, which are all SOTA neural-based news rec- ommendation methods: (1) EBNR (Okura et al., 5.3 Evaluation Metrics 2017) learns user representations by aggregating For news recommendation evaluation, we re- their browsed news with GRU. (2) DKN (Wang port the average results in terms of AUC, MRR, et al., 2018) is a deep knowledge-aware network nDCG@5 and nDCG@10. For personalized head- for news recommendation. (3) NPA (Wu et al., line generation, we evaluate the generation quality 2019b) proposes personalized attention module in using F1 ROUGE (Lin, 2004) 4 including unigram both news and user encoder. (4) NRMS (Wu et al., 4 We compute all ROUGE scores with pa- 2019c) conducts neural news recommendation with rameters “-a -c 95 -m -n 4 -w 1.2.” Refer to 87
Table 2: The overall performance of compared methods. “R-1, -2, -L” indicate F scores of ROUGE-1, -2, and -L, and “NA” denotes “Not Available”. “IM” means injection methods, c.f. 1 , 2 , and 3 in Fig. 3 for details. Metrics Methods AUC MRR NDCG@5 NDCG@10 IM ROUGE-1 ROUGE-2 ROUGE-L Pointer-Gen NA NA NA NA NA 19.86 7.76 18.83 PG+RL-ROUGE NA NA NA NA NA 20.56 8.42 20.03 1 25.13 9.03 20.73 EBNR 63.97 22.52 26.45 32.81 2 25.49 9.14 20.82 3 24.62 8.95 20.40 1 25.97 9.23 20.92 DKN 65.25 24.07 26.97 34.24 2 27.48 10.07 21.81 3 25.02 8.98 20.34 1 25.49 9.14 20.82 NPA 64.91 23.65 26.72 33.96 2 26.11 9.58 21.40 3 26.35 9.71 21.82 1 24.92 9.01 20.75 NRMS 64.27 23.28 26.60 33.58 2 26.15 9.37 21.03 3 25.41 9.12 20.91 1 23.71 8.73 21.13 LSTUR 62.49 22.69 24.71 32.28 2 24.10 8.82 20.73 3 23.11 8.42 20.38 1 27.49 10.14 21.62 NAML 66.18 25.51 27.56 35.17 2 28.01 10.72 22.24 3 27.25 10.01 21.40 and bigram overlap (ROUGE-1 and ROUGE-2) to that user modeling makes a difference in generat- assess informativeness, and the longest common ing personalized headlines. For instance, NAML subsequence (ROUGE-L) to measure fluency. Here achieves the best performance in news recommen- we adopt ROUGE because we care more about dation by learning news and user representations evaluating the recall of the generated results. All from multiple views, i.e., obtaining 66.18, 25.51, the reported values are the averaged results of 10 27.56 and 35.17 on AUC, MRR NDCG@5 and independently repeated runs. NDCG@10. Then injecting the user preferences learned by NAML to the proposed headline genera- 5.4 Experimental Results tor also gets the highest ROUGE scores with either Since we include six kinds of user modeling meth- way of the incorporation. We conjecture it is be- ods from personalized news recommendations and cause better user modeling methods can learn more propose three ways of injecting user interests in rich personalized information from click behaviors, our framework, we can derive 18 variants of ap- and well-learned user embeddings could strive to proaches that can generate personalized news head- generate better-personalized headlines. Third, it lines. Meanwhile, there are two headline genera- is reported that the second way of injecting user tion baselines, hence we totally have 20 methods interests gets the best performance on most of the for evaluation. The overall performance is illus- user modeling methods, e.g., EBNR, DKN and trated in Table 2, and we have the following obser- NAML. It is probably because the differentiation vations. of the attention distribution is intensified after the First, we can see that every personalized news user embedding perturbation, which then impacts headline generation method can outperform non- the word generation in the decoding process. How- personalized methods like PG. It might be that ever, it still remains a large room for explorations our proposed framework can generate personal- on better injecting user representations into the gen- ized news headlines by incorporating user inter- eration process since the second way seems to be ests. Such personalized headlines are more similar defective at some time. to the manually-written ones, which are taken as gold-standard in our evaluation. Second, we find https://pypi.python.org/pypi/pyrouge/0.1.3 88
Table 3: A case study on personalized headline generation for two different users by personalized (NAML+HG) and non-personalized (Pointer-Gen). Underlined words and colored words represent the correlated words in the manually-written headlines, clicked news, and the generated headlines, respectively. Case 1. Original Headline: Venezuelans rush to Peru before new requirements take effect Pointer-Gen: Venezuelans rush to Peru user A written headline: New requirements set to take effect causes Venezuelans to rush to Peru NAML+HG for user A: Peru has stricter entry requirements for escaping Venezuelans on that influx. Clicked News of user A: 1. Peru and Venezuela fans react after match ends in a draw 2. Uruguay v. Peru, Copa America and Gold Cup, Game threads and how to watch user B written headline: Venezuelan migrants to Peru face danger and discrimination NAML+HG for user B: Stricter entry requirements on Venezuelan migrants and refugees. Clicked News of user B: 1. Countries Accepting The Most Refugees (And Where They’re Coming From) 2. Venezuelan mothers, children in tow, rush to migrate 5.5 Case Study framework, can generate more condensed output based on the latent representation of news content. To further comprehend our task and the pro- However, the nature of text summarization methods posed framework, we demonstrate interesting cases without considering interactions between news and from two representative methods, namely one non- users renders them ineffective in our personalized personalized method Pointer-Gen (PG) and one headline generation. personalized method NAML+HG which utilizes Recently, stylized headlines generation were pro- the second user interests injection (c.f. Fig. 3). We posed to output eye-catching headlines by implicit also exhibit the manually-written headlines by the style transfer (Shen et al., 2017b; Fu et al., 2018; users and the original news headline as references. Prabhumoye et al., 2018) or style-oriented super- From the results shown in Table 3, we can ob- visions (Shu et al., 2018; Zhang et al., 2018; Xu serve that generated headline by non-personalized et al., 2019). However, either training a unified text method might omit some detailed but important style transfer model or constructing a personalized information. We believe the reason is that PG is text style transfer model for every user is infeasi- trained via supervised learning to maximize the log- ble due to the complex personalized style-related likelihood of ground-truth news headlines. While patterns and the limited personalized-oriented ex- our framework is trained via RL technique where amples. Meanwhile, these methods might suffer coverage score is considered as an indicator to en- from the risk of entering into click-bait territory. courage the generation to be more complete. In Personalized News Recommendation is also addition, the exhibited cases show that our frame- related to our problem. Among them, content- work can produce user-specific news headlines in based recommendations (Okura et al., 2017; Liu accordance with their individual interests reflected et al., 2010; Li et al., 2011; Lian et al., 2018; Wang by historical click behaviors. Meanwhile, some key et al., 2018; Wu et al., 2019a,b) perform user and phrases in the personalized-written titles success- news matching on a learned hidden space, and user fully appeared in the machine-generated headlines. representation is learned based on historical clicked news contents. It inspires us to personalize head- 6 Related Work line generator by incorporating user embeddings. Headline generation has been considered as spe- Deep models (Lian et al., 2018; Wang et al., 2018; cialized text summarization (Luo et al., 2019; Jia Wu et al., 2019b,a), recently, demonstrated signif- et al., 2020), from which both extractive (Dorr icant improvements because of their capabilities et al., 2003; Alfonseca et al., 2013) and abstrac- in representation learning on both user-side and tive summarization (Sun et al., 2015; Takase et al., news-side data. Different from the efforts on per- 2016; Tan et al., 2017; Gavrilov et al., 2019; See sonalized news recommendation, our work focuses et al., 2017) approaches prevailed for decades. on generating fascinating headlines for different Extractive methods select a subset of actual sen- users, which is orthogonal to existing work. tences in original article, which may derive inco- 7 Conclusion and Future Work herent summary (Alfonseca et al., 2013). While abstractive models, basically falling in an encoder- In this paper, we formulated the problem of per- decoder (Shen et al., 2017a; Murao et al., 2019) sonalized news headline generation. To provide an 89
offline testbed for this problem, we constructed a Bonnie Dorr, David Zajic, and Richard Schwartz. 2003. dataset named PENS from Microsoft News. The Hedge trimmer: A parse-and-trim approach to head- line generation. In Proceedings of the HLT-NAACL news corpus of this dataset contains more than 100 03 on Text Summarization Workshop - Volume 5, thousand news articles over 15 topic categories. HLT-NAACL-DUC ’03, page 1–8, USA. Associa- The training set constitutes of 500, 000 impressions tion for Computational Linguistics. of 445, 765 users to learn user interests and con- struct personalized news headline generator by dis- Zhenxin Fu, Xiaoye Tan, Nanyun Peng, Dongyan Zhao, and Rui Yan. 2018. Style transfer in text: Explo- tant supervisions. The test set was constructed by ration and evaluation. In AAAI, pages 663–670. 103 annotators with their clicked behaviors and manually-written personalized news headlines. We Daniil Gavrilov, Pavel Kalaidin, and Valentin Malykh. propose a generic framework that injects user inter- 2019. Self-attentive model for headline generation. In ECIR, pages 87–93. ests into an encoder-decoder headline generator in three different manners to resolve our problem. We Xiaotao Gu, Yuning Mao, Jiawei Han, Jialu Liu, You compared both SOTA user modeling and headline Wu, Cong Yu, Daniel Finnie, Hongkun Yu, Jiaqi generating approaches to present benchmark scores Zhai, and Nicholas Zukoski. 2020. Generating rep- resentative headlines for news stories. In WWW, on the proposed dataset. pages 1773–1784. For future work, we first believe designing more complex and refined approaches to generated more Ruipeng Jia, Yanan Cao, Hengzhu Tang, Fang Fang, diversified personalized news headlines will be in- Cong Cao, and Shi Wang. 2020. Neural extractive summarization with hierarchical attentive heteroge- teresting. More importantly, how to improve per- neous graph network. In Proceedings of the 2020 sonalization while keeping factualness will be an- Conference on Empirical Methods in Natural Lan- other interesting work, and it will propel the meth- guage Processing (EMNLP), pages 3622–3631, On- ods deployable in practical scenarios. Third, news line. Association for Computational Linguistics. headline personalization might burgeon the news Diederik P Kingma and Jimmy Ba. 2014. Adam: A content personalization, which is a more challeng- method for stochastic optimization. arXiv preprint ing but interesting open problem. arXiv:1412.6980. Acknowledgments Paul LaRocque. 2003. Heads you win: An easy guide to better headline and caption writing. Marion The research work is supported by the National Street Press, Inc. Key Research and Development Program of China Lei Li, Dingding Wang, Tao Li, Daniel Knox, and Bal- under Grant No. 2017YFB1002104, the National aji Padmanabhan. 2011. Scene: a scalable two-stage Natural Science Foundation of China under Grant personalized news recommendation system. In SI- No. 92046003, 61976204, U1811461. Xiang Ao GIR, pages 125–134. is also supported by the Project of Youth Innova- Lihong Li, Wei Chu, John Langford, and Robert E tion Promotion Association CAS and Beijing Nova Schapire. 2010. A contextual-bandit approach to Program Z201100006820062. personalized news article recommendation. In WWW, pages 661–670. References Jianxun Lian, Fuzheng Zhang, Xing Xie, and Guangzhong Sun. 2018. Towards better represen- Enrique Alfonseca, Daniele Pighin, and Guillermo Gar- tation learning for personalized news recommenda- rido. 2013. HEADY: News headline abstraction tion: a multi-channel deep fusion approach. In IJ- through event pattern clustering. In Proceedings CAI, pages 3805–3811. of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Chin-Yew Lin. 2004. Rouge: A package for automatic pages 1243–1253, Sofia, Bulgaria. Association for evaluation of summaries. In Text summarization Computational Linguistics. branches out, pages 74–81. Mingxiao An, Fangzhao Wu, Chuhan Wu, Kun Zhang, Jiahui Liu, Peter Dolan, and Elin Rønby Pedersen. Zheng Liu, and Xing Xie. 2019. Neural news rec- 2010. Personalized news recommendation based on ommendation with long- and short-term user repre- click behavior. In IUI, pages 31–40. sentations. In Proceedings of the 57th Annual Meet- ing of the Association for Computational Linguis- Konstantin Lopyrev. 2015. Generating news head- tics, pages 336–345, Florence, Italy. Association for lines with recurrent neural networks. arXiv preprint Computational Linguistics. arXiv:1512.01712. 90
Ling Luo, Xiang Ao, Yan Song, Feiyang Pan, Min of the AAAI Conference on Artificial Intelligence, Yang, and Qing He. 2019. Reading like HER: Hu- volume 34, pages 8910–8917. man reading inspired extractive summarization. In Proceedings of the 2019 Conference on Empirical Rui Sun, Yue Zhang, Meishan Zhang, and Donghong Methods in Natural Language Processing and the Ji. 2015. Event-driven headline generation. In Pro- 9th International Joint Conference on Natural Lan- ceedings of the 53rd Annual Meeting of the Associa- guage Processing (EMNLP-IJCNLP), pages 3033– tion for Computational Linguistics and the 7th Inter- 3043, Hong Kong, China. Association for Computa- national Joint Conference on Natural Language Pro- tional Linguistics. cessing (Volume 1: Long Papers), pages 462–472, Beijing, China. Association for Computational Lin- Kazuma Murao, Ken Kobayashi, Hayato Kobayashi, guistics. Taichi Yatsuka, Takeshi Masuyama, Tatsuru Hig- urashi, and Yoshimune Tabuchi. 2019. A case study Sho Takase, Jun Suzuki, Naoaki Okazaki, Tsutomu on neural headline generation for editing support. In Hirao, and Masaaki Nagata. 2016. Neural head- Proceedings of the 2019 Conference of the North line generation on Abstract Meaning Representation. American Chapter of the Association for Computa- In Proceedings of the 2016 Conference on Empiri- tional Linguistics: Human Language Technologies, cal Methods in Natural Language Processing, pages Volume 2 (Industry Papers), pages 73–82, Minneapo- 1054–1059, Austin, Texas. Association for Compu- lis, Minnesota. Association for Computational Lin- tational Linguistics. guistics. Jiwei Tan, Xiaojun Wan, and Jianguo Xiao. 2017. Shumpei Okura, Yukihiro Tagami, Shingo Ono, and From neural sentence summarization to headline Akira Tajima. 2017. Embedding-based news rec- generation: A coarse-to-fine approach. In IJCAI, ommendation for millions of users. In KDD, pages pages 4109–4115. 1933–1942. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Jeffrey Pennington, Richard Socher, and Christopher Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Manning. 2014. GloVe: Global vectors for word Kaiser, and Illia Polosukhin. 2017. Attention is all representation. In Proceedings of the 2014 Confer- you need. In Advances in neural information pro- ence on Empirical Methods in Natural Language cessing systems, pages 5998–6008. Processing (EMNLP), pages 1532–1543, Doha, Qatar. Association for Computational Linguistics. Michael W Wagner and Mike Gruszczynski. 2016. When framing matters: How partisan and journalis- Shrimai Prabhumoye, Yulia Tsvetkov, Ruslan Salakhut- tic frames affect individual opinions and party iden- dinov, and Alan W Black. 2018. Style transfer tification. Journalism & Communication Mono- through back-translation. In Proceedings of the graphs, 18(1):5–48. 56th Annual Meeting of the Association for Com- putational Linguistics (Volume 1: Long Papers), Hongwei Wang, Fuzheng Zhang, Xing Xie, and Minyi pages 866–876, Melbourne, Australia. Association Guo. 2018. Dkn: Deep knowledge-aware network for Computational Linguistics. for news recommendation. In WWW, pages 1835– 1844. Abigail See, Peter J. Liu, and Christopher D. Manning. 2017. Get to the point: Summarization with pointer- Chuhan Wu, Fangzhao Wu, Mingxiao An, Jianqiang generator networks. In Proceedings of the 55th An- Huang, Yongfeng Huang, and Xing Xie. 2019a. nual Meeting of the Association for Computational Neural news recommendation with attentive multi- Linguistics (Volume 1: Long Papers), pages 1073– view learning. In IJCAI, pages 3863–3869. 1083, Vancouver, Canada. Association for Computa- tional Linguistics. Chuhan Wu, Fangzhao Wu, Mingxiao An, Jianqiang Huang, Yongfeng Huang, and Xing Xie. 2019b. Shi-Qi Shen, Yan-Kai Lin, Cun-Chao Tu, Yu Zhao, Zhi- Npa: Neural news recommendation with personal- Yuan Liu, Mao-Song Sun, et al. 2017a. Recent ad- ized attention. In KDD, pages 2576–2584. vances on neural headline generation. Journal of Computer Science and Technology, 32(4):768–784. Chuhan Wu, Fangzhao Wu, Suyu Ge, Tao Qi, Tianxiao Shen, Tao Lei, Regina Barzilay, and Tommi Yongfeng Huang, and Xing Xie. 2019c. Neu- Jaakkola. 2017b. Style transfer from non-parallel ral news recommendation with multi-head self- text by cross-alignment. In NIPS, pages 6830–6841. attention. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing Kai Shu, Suhang Wang, Thai Le, Dongwon Lee, and and the 9th International Joint Conference on Natu- Huan Liu. 2018. Deep headline generation for click- ral Language Processing (EMNLP-IJCNLP), pages bait detection. In ICDM, pages 467–476. 6389–6394, Hong Kong, China. Association for Computational Linguistics. Yun-Zhu Song, Hong-Han Shuai, Sung-Lin Yeh, Yi- Lun Wu, Lun-Wei Ku, and Wen-Chih Peng. 2020. Fangzhao Wu, Ying Qiao, Jiun-Hung Chen, Chuhan Attractive or faithful? popularity-reinforced learn- Wu, Tao Qi, Jianxun Lian, Danyang Liu, Xing Xie, ing for inspired headline generation. In Proceedings Jianfeng Gao, Winnie Wu, and Ming Zhou. 2020. 91
MIND: A large-scale dataset for news recommen- dation. In Proceedings of the 58th Annual Meet- ing of the Association for Computational Linguistics, pages 3597–3606, Online. Association for Computa- tional Linguistics. Peng Xu, Chien-Sheng Wu, Andrea Madotto, and Pas- cale Fung. 2019. Clickbait? sensational headline generation with auto-tuned reinforcement learning. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Lan- guage Processing (EMNLP-IJCNLP), pages 3065– 3075, Hong Kong, China. Association for Computa- tional Linguistics. Ruqing Zhang, Jiafeng Guo, Yixing Fan, Yanyan Lan, Jun Xu, Huanhuan Cao, and Xueqi Cheng. 2018. Question headline generation for news articles. In CIKM, pages 617–626. Guanjie Zheng, Fuzheng Zhang, Zihan Zheng, Yang Xiang, Nicholas Jing Yuan, Xing Xie, and Zhen- hui Li. 2018. Drn: A deep reinforcement learn- ing framework for news recommendation. In WWW, pages 167–176. 92
You can also read