Screenplay Summarization Using Latent Narrative Structure
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
Screenplay Summarization Using Latent Narrative Structure
Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2 Mirella Lapata1
1 Institute for Language, Cognition and Computation
School of Informatics, University of Edinburgh
2 School of Computing and Information Systems
University of Melbourne
p.papalampidi@sms.ed.ac.uk, keller@inf.ed.ac.uk,
lea.frermann@unimelb.edu.au, mlap@inf.ed.ac.uk
Abstract Victim: Mike Kimble, found in a Body Farm. Died 6
Setup
hours ago, unknown cause of death.
CSI discover cow tissue in Mike's body. Opportunity
Most general-purpose extractive summariza- Cross-contamination is suggested. Probable
tion models are trained on news articles, which cause of death: Mike's house has been set on
New
fire. CSI finds blood: Mike was murdered, fire was
are short and present all important information Situation
a cover up. First suspects: Mike's fiance, Jane
upfront. As a result, such models are biased by and her ex-husband, Russ.
CSI finds photos in Mike's house of Jane's Change of
position and often perform a smart selection daughter, Jodie, posing naked. Plans
of sentences from the beginning of the doc- Mike is now a suspect of abusing Jodie. Russ
Progress
allows CSI to examine his gun.
ument. When summarizing long narratives, CSI discovers that the bullet that killed Mike
Point of
which have complex structure and present in- was made of frozen beef that melt inside him.
no Return
They also find beef in Russ' gun.
formation piecemeal, simple position heuris- Russ confesses that he knew that Mike was
tics are not sufficient. In this paper, we pro- Complications
abusing Jody, so he confronted and killed him.
pose to explicitly incorporate the underlying Russ is given bail, since no jury would convict Major
a protective father. Setback
structure of narratives into general unsuper- CSI discovers that the naked photos were taken
The final push
vised and supervised extractive summarization on a boat, which belongs to Russ.
CSI discovers that it was Russ who was
models. We formalize narrative structure in abusing his daughter based on fluids found in
Climax
terms of key narrative events (turning points) his sleeping bag and later killed Mike who tried
to help Jodie.
and treat it as latent in order to summarize Russ receives a mandatory life sentence. Aftermath
screenplays (i.e., extract an optimal sequence
of scenes). Experimental results on the CSI Figure 1: Example of narrative structure for episode
corpus of TV screenplays, which we augment “Burden of Proof” from TV series Crime Scene Inves-
with scene-level summarization labels, show tigation (CSI); turning points are highlighted in color.
that latent turning points correlate with im-
portant aspects of a CSI episode and improve
summarization performance over general ex- ements of a story in the beginning and support-
tractive algorithms, leading to more complete ing material and secondary details afterwards. The
and diverse summaries. rigid structure of news articles is expedient since
important passages can be identified in predictable
1 Introduction locations (e.g., by performing a “smart selection”
Automatic summarization has enjoyed renewed of sentences from the beginning of the document)
interest in recent years thanks to the popular- and the structure itself can be explicitly taken into
ity of modern neural network-based approaches account in model design (e.g., by encoding the rel-
(Cheng and Lapata, 2016; Nallapati et al., 2016, ative and absolute position of each sentence).
2017; Zheng and Lapata, 2019) and the avail- In this paper we are interested in summarizing
ability of large-scale datasets containing hundreds longer narratives, i.e., screenplays, whose form
of thousands of document–summary pairs (Sand- and structure is far removed from newspaper ar-
haus, 2008; Hermann et al., 2015; Grusky et al., ticles. Screenplays are typically between 110 and
2018; Narayan et al., 2018; Fabbri et al., 2019; Liu 120 pages long (20k words), their content is bro-
and Lapata, 2019). Most efforts to date have con- ken down into scenes, which contain mostly dia-
centrated on the summarization of news articles logue (lines the actors speak) as well as descrip-
which tend to be relatively short and formulaic tions explaining what the camera sees. Moreover,
following an “inverted pyramid” structure which screenplays are characterized by an underlying
places the most essential, novel and interesting el- narrative structure, a sequence of events by which
1920
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 1920–1933
July 5 - 10, 2020. c 2020 Association for Computational LinguisticsScreenplay Latent Narrative Structure
ermann et al., 2018) which revolves around a team
TP1: Introduction of forensic investigators solving criminal cases.
TP2: Goal definition
Such programs have a complex but well-defined
structure: they open with a crime, the crime scene
TP3: Commitment
is examined, the victim is identified, suspects are
TP4: Setback introduced, forensic clues are gathered, suspects
TP5: Ending are investigated, and finally the case is solved.
In this work, we adapt general-purpose extrac-
tive summarization algorithms (Nallapati et al.,
Summary scenes 2017; Zheng and Lapata, 2019) to identify infor-
Video summary
relevant mative scenes in screenplays and instill in them
to TP2
knowledge about narrative film structure (Hauge,
2017; Cutting, 2016; Freytag, 1896). Specifically,
we adopt a scheme commonly used by screen-
irrelevant writers as a practical guide for producing success-
ful screenplays. According to this scheme, well-
relevant
to TP5
structured stories consist of six basic stages which
are defined by five turning points (TPs), i.e., events
which change the direction of the narrative, and
Figure 2: We first identify scenes that act as turning determine the story’s progression and basic the-
points (i.e., key events that segment the story into sec- matic units. In Figure 1, TPs are highlighted for
tions). We next create a summary by selecting informa- a CSI episode. Although the link between turning
tive scenes, i.e.,semantically related to turning points. points and summarization has not been previously
made, earlier work has emphasized the importance
a story is defined (Cutting, 2016), and by the of narrative structure for summarizing books (Mi-
story’s characters and their roles (Propp, 1968). halcea and Ceylan, 2007) and social media content
Contrary to news articles, the gist of the story in a (Kim and Monroy-Hernández, 2015). More re-
screenplay is not disclosed at the start, information cently, Papalampidi et al. (2019) have shown how
is often revealed piecemeal; characters evolve and to identify turning points in feature-length screen-
their actions might seem more or less important plays by projecting synopsis-level annotations.
over the course of the narrative. From a modeling Crucially, our method does not involve man-
perspective, obtaining training data is particularly ually annotating turning points in CSI episodes.
problematic: even if one could assemble screen- Instead, we approximate narrative structure au-
plays and corresponding summaries (e.g., by min- tomatically by pretraining on the annotations of
ing IMDb or Wikipedia), the size of such a corpus the TRIPOD dataset of Papalampidi et al. (2019)
would be at best in the range of a few hundred and employing a variant of their model. We find
examples not hundreds of thousands. Also note that narrative structure representations learned on
that genre differences might render transfer learn- their dataset (which was created for feature-length
ing (Pan and Yang, 2010) difficult, e.g., a model films), transfer well across cinematic genres and
trained on movie screenplays might not generalize computational tasks. We propose a framework for
to sitcoms or soap operas. end-to-end training in which narrative structure is
Given the above challenges, we introduce a treated as a latent variable for summarization. We
number of assumptions to make the task feasible. extend the CSI dataset (Frermann et al., 2018) with
Firstly, our goal is to produce informative sum- binary labels indicating whether a scene should be
maries, which serve as a surrogate to reading the included in the summary and present experiments
full script or watching the entire film. Secondly, with both supervised and unsupervised summa-
we follow Gorinski and Lapata (2015) in con- rization models. An overview of our approach is
ceptualizing screenplay summarization as the task shown in Figure 2.
of identifying a sequence of informative scenes. Our contributions can be summarized as fol-
Thirdly, we focus on summarizing television pro- lows: (a) we develop methods for instilling knowl-
grams such as CSI: Crime Scene Investigation (Fr- edge about narrative structure into generic su-
1921pervised and unsupervised summarization algo- search on narrative structure analysis (Cutting,
rithms; (b) we provide a new layer of annotations 2016; Hauge, 2017), screenplay summarization
for the CSI corpus, which can be used for research (Gorinski and Lapata, 2015), and neural network
in long-form summarization; and (c) we demon- modeling (Dong, 2018). We focus on extractive
strate that narrative structure can facilitate screen- summarization and our goal is to identify an op-
play summarization; our analysis shows that key timal sequence of key events in a narrative. We
events identified in the latent space correlate with aim to create summaries which re-tell the plot of a
important summary content. story in a concise manner. Inspired by recent neu-
ral network-based approaches (Cheng and Lapata,
2 Related Work 2016; Nallapati et al., 2017; Zhou et al., 2018;
Zheng and Lapata, 2019), we develop supervised
A large body of previous work has focused on the and unsupervised models for our summarization
computational analysis of narratives (Mani, 2012; task based on neural representations of scenes
Richards et al., 2009). Attempts to analyze how and how these relate to the screenplay’s narra-
stories are written have been based on sequences tive structure. Contrary to most previous work
of events (Schank and Abelson, 1975; Chambers which has focused on characters, we select sum-
and Jurafsky, 2009), plot units (McIntyre and Lap- mary scenes based on events and their importance
ata, 2010; Goyal et al., 2010; Finlayson, 2012) and in the story. Our definition of narrative structure
their structure (Lehnert, 1981; Rumelhart, 1980), closely follows Papalampidi et al. (2019). How-
as well as on characters or personas in a narrative ever, the model architectures we propose are gen-
(Black and Wilensky, 1979; Propp, 1968; Bam- eral and could be adapted to different plot analysis
man et al., 2014, 2013; Valls-Vargas et al., 2014) schemes (Field, 2005; Vogler, 2007). To overcome
and their relationships (Elson et al., 2010; Agarwal the difficulties in evaluating summaries for longer
et al., 2014; Srivastava et al., 2016). narratives, we also release a corpus of screenplays
As mentioned earlier, work on summarization with scenes labeled as important (summary wor-
of narratives has had limited appeal, possibly due thy). Our annotations augment an existing dataset
to the lack of annotated data for modeling and based on CSI episodes (Frermann et al., 2018),
evaluation. Kazantseva and Szpakowicz (2010) which was originally developed for incremental
summarize short stories based on importance cri- natural language understanding.
teria (e.g., whether a segment contains protagonist
or location information); they create summaries to 3 Problem Formulation
help readers decide whether they are interested in
reading the whole story, without revealing its plot. Let D denote a screenplay consisting of a se-
Mihalcea and Ceylan (2007) summarize books quence of scenes D = {s1 , s2 , . . . , sn }. Our aim is
with an unsupervised graph-based approach op- to select a subset D 0 = {si , . . . , sk } consisting of
erating over segments (i.e., topical units). Their the most informative scenes (where k < n). Note
algorithm first generates a summary for each seg- that this definition produces extractive summaries;
ment and then an overall summary by collecting we further assume that selected scenes are pre-
sentences from the individual segment summaries. sented according to their order in the screenplay.
Focusing on screenplays, Gorinski and Lapata We next discuss how summaries can be created us-
(2015) generate a summary by extracting an opti- ing both unsupervised and supervised approaches,
mal chain of scenes via a graph-based approach and then move on to explain how these are adapted
centered around the main characters. In a sim- to incorporate narrative structure.
ilar fashion, Tsoneva et al. (2007) create video
summaries for TV series episodes; their algorithm 3.1 Unsupervised Screenplay Summarization
ranks sub-scenes in terms of importance using fea- Our unsupervised model is based on an extension
tures based on character graphs and textual cues of T EXT R ANK (Mihalcea and Tarau, 2004; Zheng
available in the subtitles and movie scripts. Vicol and Lapata, 2019), a well-known algorithm for ex-
et al. (2018) introduce the MovieGraphs dataset, tractive single-document summarization. In our
which also uses character-centered graphs to de- setting, a screenplay is represented as a graph, in
scribe the content of movie video clips. which nodes correspond to scenes and edges be-
Our work synthesizes various strands of re- tween scenes si and s j are weighted by their simi-
1922larity ei j . A node’s centrality (importance) is mea- We encode the screenplay with a BiLSTM net-
sured by computing its degree: work and obtain contextualized representations s0i
for scenes si by concatenating the hidden layers of
centrality(si ) = λ1 ∑ ei j + λ2 ∑ ei j (1) →
− ←
−
the forward hi and backward hi LSTM, respec-
ji →
− ←−
tively: s0i = [ hi ; hi ]. The vector s0i therefore repre-
where λ1 +λ2 = 1. The modification introduced in sents the content of the ith scene.
Zheng and Lapata (2019) takes directed edges into We also estimate the salience of scene si by
account, capturing the intuition that the centrality measuring its similarity with a global screenplay
of any two nodes is influenced by their relative po- content representation d. The latter is the weighted
sition. Also note that the edges of preceding and sum of all scene representations s1 , s2 , . . . , sn . We
following scenes are differentially weighted by λ1 calculate the semantic similarity between s0i and d
and λ2 . by computing the element-wise dot product bi , co-
Although earlier implementations of T EXT - sine similarity ci , and pairwise distance ui between
R ANK (Mihalcea and Tarau, 2004) compute node their respective vectors:
similarity based on symbolic representations such
s0i · d
as tf*idf, we adopt a neural approach. Specifically, bi = s0i d ci = (2)
s0i kdk
we obtain sentence representations based on a pre-
trained encoder. In our experiments, we rely on s0i · d
ui = (3)
the Universal Sentence Encoder (USE; Cer et al. max(ks0i k2 · kdk2 )
2018), however, other embeddings are possible.1
We represent a scene by the mean of its sentence The salience vi of scene si is the concatenation of
representations and measure scene similarity ei j the similarity metrics: vi = [bi ; ci ; ui ]. The content
using cosine.2 As in the original T EXT R ANK al- vector s0i and the salience vector vi are concate-
gorithm (Mihalcea and Tarau, 2004), scenes are nated and fed to a single neuron that outputs the
ranked based on their centrality and the M most probability of a scene belonging to the summary.3
central ones are selected to appear in the summary. 3.3 Narrative Structure
3.2 Supervised Screenplay Summarization We now explain how to inject knowledge about
narrative structure into our summarization models.
Most extractive models frame summarization as
For both models, such knowledge is transferred
a classification problem. Following a recent ap-
via a network pre-trained on the TRIPOD4 dataset
proach (S UMMA RU NN ER; Nallapati et al. 2017),
introduced by Papalampidi et al. (2019). This
we use a neural network-based encoder to build
dataset contains 99 movies annotated with turning
representations for scenes and apply a binary clas-
points. TPs are key events in a narrative that define
sifier over these to predict whether they should
the progression of the plot and occur between con-
be in the summary. For each scene si ∈ D , we
secutive acts (thematic units). It is often assumed
predict a label yi ∈ {0, 1} (where 1 means that
(Cutting, 2016) that there are six acts in a film
si must be in the summary) and assign a score
(Figure 1), each delineated by a turning point (ar-
p(yi |si , D , θ) quantifying si ’s relevance to the sum-
rows in the figure). Each of the five TPs has also a
mary (θ denotes model parameters). We assem-
well-defined function in the narrative: we present
ble a summary by selecting M sentences with the
each TP alongside with its definition as stated in
top p(1|si , D , θ).
screenwriting theory (Hauge, 2017) and adopted
We calculate sentence representations via the
by Papalampidi et al. (2019) in Table 1 (see Ap-
pre-trained USE encoder (Cer et al., 2018); a scene
pendix A for a more detailed description of nar-
is represented as the weighted sum of the repre-
rative structure theory). Papalampidi et al. (2019)
sentations of its sentences, which we obtain from
identify scenes in movies that correspond to these
a BiLSTM equipped with an attention mechanism.
key events as a means for analyzing the narrative
Next, we compute richer scene representations by
modeling surrounding context of a given scene. 3 Aside from salience and content, Nallapati et al. (2017)
take into account novelty and position-related features. We
1 USE performed better than BERT in our experiments. ignore these as they are specific to news articles and denote
2 We found cosine to be particularly effective with USE the modified model as S UMMA RU NN ER *.
representations; other metrics are also possible. 4 https://github.com/ppapalampidi/TRIPOD
1923Turning Point Definition Summarization with Narrative Structure).5 We
Introductory event that occurs after
TP1: Opportunity the presentation of the story setting. first present an unsupervised variant which mod-
Event where the main goal of the ifies the computation of scene centrality in the di-
TP2: Change of Plans story is defined.
rected version of T EXT R ANK (Equation (1)).
Event that pushes the main charac-
TP3: Point of No Return ter(s) to fully commit to their goal. Specifically, we use the pre-trained network de-
Event where everything falls apart scribed in Section 3.3 to obtain TP-specific at-
TP4: Major Setback (temporarily or permanently). tention distributions. We then select an overall
Final event of the main story, mo-
TP5: Climax ment of resolution. score fi for each scene (denoting how likely it is
to act as a TP). We set fi = max j∈[1,5] pi j , i.e., to
Table 1: Turning points and their definitions as given
in Papalampidi et al. (2019) the pi j value that is highest across TPs. We incor-
porate these scores into centrality as follows:
structure of movies. They collect sentence-level centrality(si )=λ1 ∑ (ei j + f j )+λ2 ∑ (ei j + fi ) (6)
TP annotations for plot synopses and subsequently ji
project them via distant supervision onto screen-
plays, thereby creating silver-standard labels. We Intuitively, we add the f j term in the forward sum
utilize this silver-standard dataset in order to pre- in order to incrementally increase the centrality
train a network which performs TP identification. scores of scenes as the story moves on and we en-
counter more TP events (i.e., we move to later sec-
TP Identification Network We first encode tions in the narrative). At the same time, we add
screenplay scenes via a BiLSTM equipped with an the fi term in the backward sum in order to also
attention mechanism. We then contextualize them increase the scores of scenes identified as TPs.
with respect to the whole screenplay via a second
Supervised S UMMER We also propose a su-
BiLSTM. Next, we compute topic-aware scene
pervised variant of S UMMER following the basic
representations ti via a context interaction layer
model formulation in Section 3.3. We still repre-
(CIL) as proposed in Papalampidi et al. (2019).
sent a scene as the concatenation of a content vec-
CIL is inspired by traditional segmentation ap-
tor s0 and salience vector v0 , which serve as input
proaches (Hearst, 1997) and measures the seman-
to a binary classifier. However, we now modify
tic similarity of the current scene with a preceding
how salience is determined; instead of comput-
and following context window in the screenplay.
ing a general global content representation d for
Hence, the topic-aware scene representations also
the screenplay, we identify a sequence of TPs and
encode the degree to which each scene acts as a
measure the semantic similarity of each scene with
topic boundary in the screenplay.
this sequence. Our model is depicted in Figure 3.
In the final layer, we employ TP-specific atten-
We utilize the pre-trained TP network (Fig-
tion mechanisms to compute the probability pi j
ures 3(a) and (b)) to compute sparse attention
that scene ti represents the jth TP in the screen-
scores over scenes. In the supervised setting,
play. Note that we expect the TP-specific atten-
where gold-standard binary labels provide a train-
tion distributions to be sparse, as there are only
ing signal, we fine-tune the network in an end-to-
a few scenes which are relevant for a TP (recall
end fashion on summarization (Figure 3(c)). We
that TPs are boundary scenes between sections).
compute the TP representations via the attention
To encourage sparsity, we add a low temperature
scores; we calculate a vector t p j as the weighted
value τ (Hinton et al., 2015) to the softmax part of
sum of all topic-aware scene representations t pro-
the attention mechanisms:
duced via CIL: t p j = ∑i∈[1,N] pi j ti , where N is the
gi j = tanh(W j ti + b j ), g j ∈ [−1, 1] (4) number of scenes in a screenplay. In practice, only
T
a few scenes contribute to t p j due to the τ param-
exp(gi j /τ) eter in the softmax function (Equation (5)).
pi j = T , ∑ pi j = 1 (5)
∑t=1 exp(gt j /τ) i=1 A TP-scene interaction layer measures the se-
mantic similarity between scenes ti and latent TP
where W j , b j represent the trainable weights of the
representations t p j (Figure 3(c)). Intuitively, a
attention layer of the jth TP.
complete summary should contain scenes which
Unsupervised S UMMER We now introduce 5 We make our code publicly available at https://
our model, S UMMER (short for Screenplay github.com/ppapalampidi/SUMMER.
1924(c): Summary scenes prediction when a TP should occur (e.g., the Opportunity oc-
Final scene
curs after the first 10% of a screenplay, Change of
..
y1, ..., yk, ..., yM
representations Plans is approximately 25% in). It is reasonable to
discourage t p distributions to deviate drastically
TP-scene from these expected positions. Focal regulariza-
..
..
Interaction
Layer tion F minimizes the KL divergence DKL between
Salience wrt Content each TP attention distribution t pi and its expected
plotline
position distribution thi :
(b): Narrative structure prediction
tp1 tp2 tp3 tp4 tp5 F= ∑ DKL (t pi kthi ) (8)
i∈[1,5]
The final loss L is the weighted sum of all three
TP1 TP2 TP3 TP4 TP5
components, where a, b are fixed during training:
L = BCE + aO + bF.
t1 ... tk ... tM
4 Experimental Setup
Content Interaction Layer
Crime Scene Investigation Dataset We per-
s'1 ... s'k ... s'M
formed experiments on an extension of the CSI
Screenplay Encoder
(BiLSTM)
dataset6 introduced by Frermann et al. (2018). It
consists of 39 CSI episodes, each annotated with
s1 ... sk ... sM
(a): Scene encoding word-level labels denoting whether the perpetra-
tor is mentioned in the utterances characters speak.
Figure 3: Overview of S UMMER. We use one
We further collected scene-level binary labels in-
TP-specific attention mechanism per turning point in
order to acquire TP-specific distributions over scenes. dicating whether episode scenes are important and
We then compute the similarity between TPs and con- should be included in a summary. Three human
textualized scene representations. Finally, we perform judges performed the annotation task after watch-
max pooling over TP-specific similarity vectors and ing the CSI episodes scene-by-scene. To facilitate
concatenate the final similarity representation with the the annotation, judges were asked to indicate why
contextualized scene representation. they thought a scene was important, citing the fol-
lowing reasons: it revealed (i) the victim, (ii) the
are related to at least one of the key events in cause of death, (iii) an autopsy report, (iv) crucial
the screenplay. We calculate the semantic similar- evidence, (v) the perpetrator, and (vi) the motive or
ity vi j of scene ti with TP t p j as in Equations (2) the relation between perpetrator and victim. An-
and (3). We then perform max pooling over vec- notators were free to select more than one or none
tors vi1 , . . . , viT , where T is the number of TPs of the listed reasons where appropriate. We can
(i.e., five) and calculate a final similarity vector v0i think of these reasons as high-level aspects a good
for the ith scene. summary should cover (for CSI and related crime
The model is trained end-to-end on the summa- series). Annotators were not given any informa-
rization task using BCE, the binary cross-entropy tion about TPs or narrative structure; the annota-
loss function. We add an extra regularization term tion was not guided by theoretical considerations,
to this objective to encourage the TP-specific at- rather our aim was to produce useful CSI sum-
tention distributions to be orthogonal (since we maries. Table 2 presents the dataset statistics (see
want each attention layer to attend to different also Appendix B for more detail).
parts of the screenplay). We thus maximize the
Implementation Details In order to set the hy-
Kullback-Leibler (KL) divergence DKL between
perparameters of all proposed networks, we used
all pairs of TP attention distributions t pi , i ∈ [1, 5]:
a small development set of four episodes from the
1 CSI dataset (see Appendix B for details). After ex-
O= ∑ ∑ log (7) perimentation, we set the temperature τ of the soft-
i∈[1,5] j∈[1,5], j6=i DKL t pi t p j + ε
max layers for the TP-specific attentions (Equa-
Furthermore, we know from screenwriting theory tion (5)) to 0.01. Since the binary labels in the
(Hauge, 2017) that there are rules of thumb as to 6 https://github.com/EdinburghNLP/csi-corpus
1925overall Model F1
episodes 39 Lead 30% 30.66
scenes 1544 Last 30% 39.85
summary scenes 454 Mixed 30% 34.32
per episode T EXT R ANK, undirected, tf*idf 32.11
scenes 39.58 (6.52) T EXT R ANK, directed, neural 41.75
crime-specific aspects 5.62 (0.24) T EXT R ANK, directed, expected TP positions 41.05
summary scenes 11.64 (2.98) S CENE S UM, directed, character-based weights 42.02
summary scenes (%) 29.75 (7.35) S UMMER 44.70
sentences 822.56 (936.23)
Table 3: Unsupervised screenplay summarization.
tokens 13.27k (14.67k)
per episode scene Coverage # scenes
sentences 20.78 (35.61) F1
of aspects per TP
tokens 335.19 (547.61) Lead 30% 30.66 – –
tokens per sentence 16.13 (16.32) Last 30% 39.85 – –
Mixed 30% 34.32 – –
Table 2: CSI dataset statistics; means and (std).
S UMMA RU NN ER * 48.56 – –
S CENE S UM 47.71 – –
S UMMER, fixed one-hot TPs 46.92 63.11 1.00
supervised setting are imbalanced, we apply class S UMMER, fixed distributions 47.64 67.01 1.05
weights to the binary cross-entropy loss of the re- S UMMER, −P, −R 51.93 44.48 1.19
spective models. We weight each class by its in- S UMMER, −P, +R 49.98 51.96 1.14
S UMMER, +P, −R 50.56 62.35 3.07
verse frequency in the training set. Finally, in su- S UMMER, +P, +R 52.00 70.25 1.20
pervised S UMMER, where we also identify the nar-
Table 4: Supervised screenplay summarization; for in
rative structure of the screenplays, we consider as S UMMER variants, we also report the percentage of as-
key events per TP the scenes that correspond to an pect labels covered by latent TP predictions.
attention score higher than 0.05. More implemen-
tation details can be found in Appendix C. tion in Gorinski and Lapata (2015):
As shown in Table 2, the gold-standard sum-
maries in our dataset have a compression rate of ∑c∈C [c ∈ S ∪ main(C)]
ci = (9)
approximately 30%. During inference, we select ∑c∈C [c ∈ S]
the top M scenes as the summary, such that they
correspond to 30% of the length of the episode. where S is the set of all characters participating in
scene si , C is the set of all characters participat-
5 Results and Analysis ing in the screenplay and main(C) are all the main
characters of the screenplay. We retrieve the set
Is Narrative Structure Helpful? We perform of main characters from the IMDb page of the re-
10-fold cross-validation and evaluate model per- spective episode. We also note that human agree-
formance in terms of F1 score. Table 3 sum- ment for our task is 79.26 F1 score, as measured
marizes the results of unsupervised models. We on a small subset of the corpus.
present the following baselines: Lead 30% se- As shown in Table 3, S UMMER achieves the
lects the first 30% of an episode as the summary, best performance (44.70 F1 score) among all mod-
Last 30% selects the last 30%, and Mixed 30%, els and is superior to an equivalent model which
randomly selects 15% of the summary from uses expected TP positions or a character-based
the first 30% of an episode and 15% from the representation. This indicates that the pre-trained
last 30%. We also compare S UMMER against network provides better predictions for key events
T EXT R ANK based on tf*idf (Mihalcea and Ta- than position and character heuristics, even though
rau, 2004), the directed neural variant described there is a domain shift from Hollywood movies
in Section 3.1 without any TP information, a in the TRIPOD corpus to episodes of a crime
variant where TPs are approximated by their ex- series in the CSI corpus. Moreover, we find
pected position as postulated in screenwriting the- that the directed versions of T EXT R ANK are bet-
ory, and a variant that incorporates information ter at identifying important scenes than the undi-
about characters (Gorinski and Lapata, 2015) in- rected version. We found that performance peaks
stead of narrative structure. For the character- with λ1 = 0.7 (see Equation (6)), indicating that
based T EXT R ANK, called S CENE S UM, we substi- higher importance is given to scenes as the story
tute the fi , f j scores in Equation (6) with character- progresses (see Appendix D for experiments with
related importance scores ci similar to the defini- different λ values).
1926In Table 4, we report results for supervised or provides important evidence). Specifically, we
models. Aside from the various baselines in the compute the extent to which these aspects overlap
first block of the table, we compare the neural with the TPs predicted by S UMMER as:
extractive model S UMMA RU NN ER *7 (Nallapati
∑Ai ∈A ∑T Pj ∈T P [dist(T Pj ,Ai )≤1]
et al., 2017) presented in Section 3.2 with sev- C= (10)
eral variants of our model S UMMER. We exper- |A|
imented with randomly initializing the network where A is the set of all aspect scenes, |A| is the
for TP identification (−P) and with using a pre- number of aspects, T P is the set of scenes inferred
trained network (+P). We also experimented with as TPs by the model, Ai and T Pj are the subsets
removing the regularization terms, O and F (Equa- of scenes corresponding to the ith aspect and jth
tions (7) and (8)) from the loss (−R). We as- TP, respectively, and dist(T Pj , Ai ) is the minimum
sess the performance of S UMMER when we follow distance between T Pj and Ai in number of scenes.
a two-step approach where we first predict TPs The proportion of aspects covered is given in
via the pre-trained network and then train a net- Table 4, middle column. We find that coverage is
work on screenplay summarization based on fixed relatively low (44.48%) for the randomly initial-
TP representations (fixed one-hot TPs), or alter- ized S UMMER with no regularization. There is a
natively use expected TP position distributions as slight improvement of 7.48% when we force the
postulated in screenwriting theory (fixed distribu- TP-specific attention distributions to be orthogo-
tions). Finally, we incorporate character-based in- nal and close to expected positions. Pre-training
formation into our baseline and create a supervised and regularization provide a significant boost, in-
version of S CENE S UM. We now utilize the charac- creasing coverage to 70.25%, while pre-trained
ter importance scores per scene (Equation (9)) as S UMMER without regularization infers on aver-
attention scores – instead of using a trainable at- age more scenes representative of each TP. This
tention mechanism – when computing the global shows that the orthogonal constraint also encour-
screenplay representation d (Section 3.2). ages sparse attention distributions for TPs.
Table 4 shows that all end-to-end S UMMER Table 5 shows the degree of association be-
variants outperform S UMMA RU NN ER *. The tween individual TPs and summary aspects (see
best result (52.00 F1 Score) is achieved by pre- Appendix D for illustrated examples). We observe
trained S UMMER with regularization, outperform- that Opportunity and Change of Plans are mostly
ing S UMMA RU NN ER * by an absolute difference associated with information about the crime scene
of 3.44. The randomly initialized version with and the victim, Climax is focused on the revelation
no regularization achieves similar performance of the motive, while information relating to cause
(51.93 F1 score). For summarizing screenplays, of death, perpetrator, and evidence is captured by
explicitly encoding narrative structure seems to both Point of no Return and Major Setback. Over-
be more beneficial than general representations all, the generic Hollywood-inspired TP labels are
of scene importance. Finally, two-step versions adjusted to our genre and describe crime-related
of S UMMER perform poorly, which indicates that key events, even though no aspect labels were pro-
end-to-end training and fine-tuning of the TP iden- vided to our model during training.
tification network on the target dataset is crucial.
Do Humans Like the Summaries? We also
What Does the Model Learn? Apart from per- conducted a human evaluation experiment using
formance on summarization, we would also like to the summaries created for 10 CSI episodes.8 We
examine the quality of the TPs inferred by S UM - produced summaries based on the gold-standard
MER (supervised variant). Problematically, we do annotations (Gold), S UMMA RU NN ER *, and the
not have any gold-standard TP annotation in the supervised version of S UMMER. Since 30% of
CSI corpus. Nevertheless, we can implicitly assess an episode results in lengthy summaries (15 min-
whether they are meaningful by measuring how utes on average), we further increased the com-
well they correlate with the reasons annotators cite pression rate for this experiment by limiting each
to justify their decision to include a scene in the summary to six scenes. For the gold standard con-
summary (e.g., because it reveals cause of death dition, we randomly selected exactly one scene
7 Our adaptation of S UMMA RU NN ER that considers con- 8 https://github.com/ppapalampidi/SUMMER/tree/
tent and salience vectors for scene selection. master/video_summaries
1927Turning Point Crime scene Victim Death Cause Perpetrator Evidence Motive
Opportunity 56.76 52.63 15.63 15.38 2.56 0.00
Change of Plans 27.03 42.11 21.88 15.38 5.13 0.00
Point of no Return 8.11 13.16 9.38 25.64 48.72 5.88
Major Setback 0.00 0.00 6.25 10.25 48.72 35.29
Climax 2.70 0.00 6.25 2.56 23.08 55.88
Table 5: Percentage of aspect labels covered per TP for S UMMER, +P, +R.
System Crime scene Victim Death Cause Perpetrator Evidence Motive Overall Rank
S UMMA RU NN ER * 85.71 93.88 75.51 81.63 59.18 38.78 72.45 2.18
S UMMER 89.80 87.76 83.67 81.63 77.55 57.14 79.59 2.00
Gold 89.80 91.84 71.43 83.67 65.31 57.14 76.53 1.82
Table 6: Human evaluation: percentage of yes answers by AMT workers regarding each aspect in a summary. All
differences in (average) Rank are significant (p < 0.05, using a χ2 test).
per aspect. For S UMMA RU NN ER * and S UMMER can be integrated with supervised and unsuper-
we selected the top six predicted scenes based vised extractive summarization algorithms. Ex-
on their posterior probabilities. We then created periments on the CSI corpus showed that this
video summaries by isolating and merging the se- scheme transfers well to a different genre (crime
lected scenes in the raw video. investigation) and that utilizing narrative struc-
We asked Amazon Mechanical Turk (AMT) ture boosts summarization performance, leading
workers to watch the video summaries for all sys- to more complete and diverse summaries. Anal-
tems and rank them from most to least informa- ysis of model output further revealed that latent
tive. They were also presented with six questions events encapsulated by turning points correlate
relating to the aspects the summary was supposed with important aspects of a CSI summary.
to cover (e.g., Was the victim revealed in the sum- Although currently our approach relies solely
mary? Do you know who the perpetrator was?). on textual information, it would be interesting to
They could answer Yes, No, or Unsure. Five work- incorporate additional modalities such as video or
ers evaluated each summary. audio. Audiovisual information could facilitate
Table 6 shows the proportion of times partic- the identification of key events and scenes. Be-
ipants responded Yes for each aspect across the sides narrative structure, we would also like to ex-
three systems. Although S UMMER does not im- amine the role of emotional arcs (Vonnegut, 1981;
prove over S UMMA RU NN ER * in identifying ba- Reagan et al., 2016) in a screenplay. An often in-
sic information (i.e., about the victim and perpe- tegral part of a compelling story is the emotional
trator), it creates better summaries overall with experience that is evoked in the reader or viewer
more diverse content (i.e., it more frequently in- (e.g., somebody gets into trouble and then out of it,
cludes information about cause of death, evidence, somebody finds something wonderful, loses it, and
and motive). This observation validates our as- then finds it again). Understanding emotional arcs
sumption that identifying scenes that are semanti- may be useful to revealing a story’s shape, high-
cally close to the key events of a screenplay leads lighting important scenes, and tracking how the
to more complete and detailed summaries. Fi- story develops for different characters over time.
nally, Table 6 also lists the average rank per system
Acknowledgments
(lower is better), which shows that crowdwork-
ers like gold summaries best, S UMMER is often We thank the anonymous reviewers for their feed-
ranked second, followed by S UMMA RU NN ER * in back. We gratefully acknowledge the support of
third place. the European Research Council (Lapata; award
681760, “Translating Multiple Modalities into
6 Conclusions Text”) and of the Leverhulme Trust (Keller; award
IAF-2017-019).
In this paper we argued that the underlying struc-
ture of narratives is beneficial for long-form sum-
marization. We adapted a scheme for identifying References
narrative structure (i.e., turning points) in Holly- Apoorv Agarwal, Sriramkumar Balasubramanian,
wood movies and showed how this information Jiehan Zheng, and Sarthak Dash. 2014. Parsing
1928Screenplays for Extracting Social Networks from Syd Field. 2005. Screenplay: Foundations of Screen-
Movies. In Proceedings of the 3rd Workshop on writing. Dell Publishing Company.
Computational Linguistics for Literature, pages 50–
58, Gothenburg, Sweden. Mark Alan Finlayson. 2012. Learning Narrative Struc-
ture from Annotated Folktales. Ph.D. thesis, Mas-
David Bamman, Brendan O’Connor, and Noah A. sachusetts Institute of Technology.
Smith. 2013. Learning latent personas of film char-
acters. In Proceedings of the 51st Annual Meeting of Lea Frermann, Shay B Cohen, and Mirella Lapata.
the Association for Computational Linguistics (Vol- 2018. Whodunnit? crime drama as a case for natural
ume 1: Long Papers), pages 352–361, Sofia, Bul- language understanding. Transactions of the Associ-
garia. ation of Computational Linguistics, 6:1–15.
David Bamman, Ted Underwood, and Noah A. Smith. Gustav Freytag. 1896. Freytag’s technique of the
2014. A Bayesian mixed effects model of literary drama: an exposition of dramatic composition and
character. In Proceedings of the 52nd Annual Meet- art. Scholarly Press.
ing of the Association for Computational Linguis-
tics (Volume 1: Long Papers), pages 370–379, Bal- Philip John Gorinski and Mirella Lapata. 2015. Movie
timore, Maryland. script summarization as graph-based scene extrac-
tion. In Proceedings of the 2015 Conference of
John B Black and Robert Wilensky. 1979. An the North American Chapter of the Association for
evaluation of story grammars. Cognitive science, Computational Linguistics: Human Language Tech-
3(3):213–229. nologies, pages 1066–1076, Denver, Colorado. As-
sociation for Computational Linguistics.
Charles Oscar Brink. 2011. Horace on Poetry: The’Ars
Poetica’. Cambridge University Press. Amit Goyal, Ellen Riloff, and Hal Daumé III. 2010.
Automatically producing plot unit representations
Daniel Cer, Yinfei Yang, Sheng-yi Kong, Nan Hua, for narrative text. In Proceedings of the 2010 Con-
Nicole Limtiaco, Rhomni St John, Noah Constant, ference on Empirical Methods in Natural Language
Mario Guajardo-Cespedes, Steve Yuan, Chris Tar, Processing, pages 77–86, Cambridge, MA.
et al. 2018. Universal sentence encoder. arXiv
preprint arXiv:1803.11175. Max Grusky, Mor Naaman, and Yoav Artzi. 2018.
NEWSROOM: A dataset of 1.3 million summaries
Nathanael Chambers and Dan Jurafsky. 2009. Unsu- with diverse extractive strategies. In Proceedings
pervised learning of narrative schemas and their par- of the 16th Annual Conference of the North Amer-
ticipants. In Proceedings of the Joint Conference of ican Chapter of the Association for Computational
the 47th Annual Meeting of the ACL and the 4th In- Linguistics: Human Language Technologies, pages
ternational Joint Conference on Natural Language 708–719, New Orleans, USA.
Processing of the AFNLP, pages 602–610, Suntec,
Singapore. Michael Hauge. 2017. Storytelling Made Easy: Per-
suade and Transform Your Audiences, Buyers, and
Jianpeng Cheng and Mirella Lapata. 2016. Neural Clients – Simply, Quickly, and Profitably. Indie
summarization by extracting sentences and words. Books International.
In Proceedings of the 54th Annual Meeting of the
Association for Computational Linguistics (Volume Marti A Hearst. 1997. Texttiling: Segmenting text into
1: Long Papers), pages 484–494. multi-paragraph subtopic passages. Computational
linguistics, 23(1):33–64.
James E Cutting. 2016. Narrative theory and the dy-
namics of popular movies. Psychonomic bulletin & Karl Moritz Hermann, Tomáš Kočiský, Edward
review, 23(6):1713–1743. Grefenstette, Lasse Espeholt, Will Kay, Mustafa Su-
leyman, and Phil Blunsom. 2015. Teaching ma-
Yue Dong. 2018. A survey on neural network-based chines to read and comprehend. In Advances in Neu-
summarization methods. ArXiv, abs/1804.04589. ral Information Processing Systems 28, pages 1693–
1701. Morgan, Kaufmann.
David K. Elson, Nicholas Dames, and Kathleen R.
Mckeown. 2010. Extracting social networks from Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. 2015.
literary fiction. In Proceedings of the 48th Annual Distilling the knowledge in a neural network. arXiv
Meeting of the Association for Computational Lin- preprint arXiv:1503.02531.
guistics, pages 138–147, Uppsala, Sweden.
Anna Kazantseva and Stan Szpakowicz. 2010. Sum-
Alexander Fabbri, Irene Li, Tianwei She, Suyi Li, and marizing short stories. Computational Linguistics,
Dragomir Radev. 2019. Multi-news: A large-scale 36(1):71–109.
multi-document summarization dataset and abstrac-
tive hierarchical model. In Proceedings of the 57th Joy Kim and Andrés Monroy-Hernández. 2015. Sto-
Annual Meeting of the Association for Computa- ria: Summarizing social media content based on
tional Linguistics, pages 1074–1084, Florence, Italy. narrative theory using crowdsourcing. CoRR,
Association for Computational Linguistics. abs/1509.03026.
1929Diederik P Kingma and Jimmy Ba. 2014. Adam: A Adam Paszke, Sam Gross, Soumith Chintala, Gre-
method for stochastic optimization. arXiv preprint gory Chanan, Edward Yang, Zachary DeVito, Zem-
arXiv:1412.6980. ing Lin, Alban Desmaison, Luca Antiga, and Adam
Lerer. 2017. Automatic differentiation in pytorch.
Wendy G. Lehnert. 1981. Plot units and narrative sum-
marization. Cognitive Science, 5(4):293–331. Patrice Pavis. 1998. Dictionary of the theatre: Terms,
concepts, and analysis. University of Toronto Press.
Yang Liu and Mirella Lapata. 2019. Hierarchical trans-
formers for multi-document summarization. In Pro- Vladimir Iakovlevich Propp. 1968. Morphology of the
ceedings of the 57th Annual Meeting of the Asso- Folktale. University of Texas.
ciation for Computational Linguistics, pages 5070–
5081, Florence, Italy. Association for Computa- Andrew J. Reagan, Lewis Mitchell, Dilan Kiley,
tional Linguistics. Christopher M. Danforth, and Peter Sheridan Dodds.
2016. The emotional arcs of stories are dominated
Interjeet Mani. 2012. Computational Modeling of by six basic shapes. EPJ Data Science, 5(31):1–12.
Narative. Synthesis Lectures on Human Language
Technologies. Morgan and Claypool Publishers. Whitman Richards, Mark Alan Finlayson, and
Patrick Henry Winston. 2009. Advancing com-
Neil McIntyre and Mirella Lapata. 2010. Plot induc- putational models of narrative. Technical Report
tion and evolutionary search for story generation. In 63:2009, MIT Computer Science and Atrificial In-
Proceedings of the 48th Annual Meeting of the Asso- telligence Laboratory.
ciation for Computational Linguistics, pages 1562–
1572, Uppsala, Sweden. David E. Rumelhart. 1980. On evaluating story gram-
mars. Cognitive Science, 4(3):313–316.
Rada Mihalcea and Hakan Ceylan. 2007. Explorations
in automatic book summarization. In Proceedings Evan Sandhaus. 2008. The New York Times Annotated
of the 2007 joint conference on empirical methods Corpus. Linguistic Data Consortium, Philadelphia,
in natural language processing and computational 6(12).
natural language learning (EMNLP-CoNLL), pages Roger C. Schank and Robert P. Abelson. 1975. Scripts,
380–389. plans, and knowledge. In Proceedings of the 4th
Rada Mihalcea and Paul Tarau. 2004. Textrank: Bring- International Joint Conference on Artificial Intelli-
ing order into text. In Proceedings of the 2004 con- gence, pages 151–157, Tblisi, USSR.
ference on empirical methods in natural language Shashank Srivastava, Snigdha Chaturvedi, and Tom
processing, pages 404–411. Mitchell. 2016. Inferring interpersonal relations in
Ramesh Nallapati, Feifei Zhai, and Bowen Zhou. 2017. narrative summaries. In Proceedings of the 13th
Summarunner: A recurrent neural network based se- AAAI Conference on Artificial Intelligence, pages
quence model for extractive summarization of docu- 2807–2813, Phoenix, Arizona. AAAI Press.
ments. In Thirty-First AAAI Conference on Artificial Kristin Thompson. 1999. Storytelling in the new Holly-
Intelligence. wood: Understanding classical narrative technique.
Harvard University Press.
Ramesh Nallapati, Bowen Zhou, Caglar Gulcehre,
Bing Xiang, et al. 2016. Abstractive text summa- Tsvetomira Tsoneva, Mauro Barbieri, and Hans Weda.
rization using sequence-to-sequence rnns and be- 2007. Automated summarization of narrative video
yond. arXiv preprint arXiv:1602.06023. on a semantic level. In International Conference on
Semantic Computing (ICSC 2007), pages 169–176.
Shashi Narayan, Shay B. Cohen, and Mirella Lapata.
IEEE.
2018. Don’t give me the details, just the summary!
topic-aware convolutional neural networks for ex- Josep Valls-Vargas, J. Zhu, and Santiago Ontanon.
treme summarization. In Proceedings of the 2018 2014. Toward automatic role identification in unan-
Conference on Empirical Methods in Natural Lan- notated folk tales. In Proceedings of the 10th AAAI
guage Processing, pages 1797–1807, Brussels, Bel- Conference on Artificial Intelligence and Interactive
gium. Association for Computational Linguistics. Digital Entertainment, pages 188–194.
Sinno Jialin Pan and Qiang Yang. 2010. A survey on Paul Vicol, Makarand Tapaswi, Lluis Castrejon, and
transfer learning. IEEE Transactions on Knowledge Sanja Fidler. 2018. Moviegraphs: Towards under-
and Data Engineering, 22(10):1345–1359. standing human-centric situations from videos. In
Proceedings of the IEEE Conference on Computer
Pinelopi Papalampidi, Frank Keller, and Mirella La- Vision and Pattern Recognition, pages 8581–8590.
pata. 2019. Movie plot analysis via turning point
identification. In Proceedings of the 2019 Confer- Christopher Vogler. 2007. Writer’s Journey: Mythic
ence on Empirical Methods in Natural Language Structure for Writers. Michael Wiese Productions.
Processing and the 9th International Joint Confer-
ence on Natural Language Processing (EMNLP- Kurt Vonnegut. 1981. Palm Sunday. RosettaBooks
IJCNLP), pages 1707–1717. LLC, New York.
1930Crime scene Motive Directed neural
44 TextRank
42 SUMMER
12.4% 10.9%
Victim 40
14.6%
F1 score (%)
38
36.1%
14.4% Evidence 36
Perpetrator 11.6% 34
32
Cause of death
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
1
Figure 4: Average composition of a CSI summary Figure 5: F1 score (%) for directed neural T EXT -
based on different crime-related aspects. R ANK and S UMMER for unsupervised summarization
with respect to different λ1 values. Higher λ1 values
correspond to higher importance in the next context for
Hao Zheng and Mirella Lapata. 2019. Sentence cen- the centrality computation of a current scene.
trality revisited for unsupervised summarization.
arXiv preprint arXiv:1906.03508.
Qingyu Zhou, Nan Yang, Furu Wei, Shaohan Huang, B CSI Corpus
Ming Zhou, and Tiejun Zhao. 2018. Neural docu-
ment summarization by jointly learning to score and
select sentences. arXiv preprint arXiv:1807.02305.
As described in Section 4, we collected aspect-
based summary labels for all episodes in the CSI
A Narrative Structure Theory corpus. In Figure 4 we illustrate the average com-
position of a summary based on the different as-
The initial formulation of narrative structure was
pects seen in a crime investigation (e.g., crime
promoted by Aristotle, who defined the basic
scene, victim, cause of death, perpetrator, evi-
triangle-shaped plot structure, that has a beginning
dence). Most of these aspects are covered in
(protasis), middle (epitasis) and end (catastrophe)
10–15% of a summary, which corresponds to ap-
(Pavis, 1998). However, later theories argued that
proximately two scenes in the episode. Only
the structure of a play should be more complex
the “Evidence” aspect occupies a larger propor-
(Brink, 2011) and hence, other schemes (Freytag,
tion of the summary (36.1%) corresponding to
1896) were proposed with fine-grained stages and
five scenes. However, there exist scenes which
events defining the progression of the plot. These
cover multiple aspects (an as a result are anno-
events are considered as the precursor of turning
tated with more than one label) and episodes that
points, defined by Thompson (1999) and used in
do not include any scenes related to a specific as-
modern variations of screenplay theory. Turning
pect (e.g., if the murder was a suicide, there is no
points are narrative moments from which the plot
perpetrator).
goes in a different direction. By definition these
occur at the junctions of acts. We should note that Frermann et al. (2018) dis-
Currently, there are myriad schemes describ- criminate between different cases presented in the
ing the narrative structure of films, which are of- same episode in the original CSI dataset. Specif-
ten used as a practical guide for screenwriters ically, there are episodes in the dataset, where ex-
(Cutting, 2016). One variation of these modern cept for the primary crime investigation case, a
schemes is adopted by Papalampidi et al. (2019), second one is presented occupying a significantly
who focus on the definition of turning points and smaller part of the episode. Although in the origi-
demonstrate that such events indeed exist in films nal dataset, there are annotations available indicat-
and can be automatically identified. According ing which scenes refer to each case, we assume no
to the adopted scheme (Hauge, 2017), there are such knowledge treating the screenplay as a single
six stages (acts) in a film, namely the setup, the unit — most TV series and movies contain sub-
new situation, progress, complications and higher stories. We also hypothesize that the latent iden-
stakes, the final push and the aftermath, separated tified TP events in S UMMER should relate to the
by the five turning points presented in Table 1. primary case.
1931Figure 6: Examples of inferred TPs alongside with gold-standard aspect-based summary labels in CSI episodes at
test time. The TP events are identified in the latent space for the supervised version of S UMMER (+P, +R).
C Implementation Details the preceding ones, since λ1 and λ2 are bounded
(λ1 + λ2 = 1).
In all unsupervised versions of T EXT R ANK and
We observe that performance increases when
S UMMER we used a threshold h equal to 0.2 for re-
higher importance is attributed to screenplay
moving weak edges from the corresponding fully
scenes as the story moves on (λ1 > 0.5), whereas
connected screenplay graphs. For the supervised
for extreme cases (λ1 → 1), where only the later
version of S UMMER, where we use additional reg-
parts of the story are considered, performance
ularization terms in the loss function, we experi-
drops. Overall, the same peak appears for both
mentally set the weights a and b for the different
T EXT R ANK and S UMMER when λ1 ∈ [0.6, 0.7],
terms to 0.15 and 0.1, respectively.
which means that slightly higher importance is at-
We used the Adam algorithm (Kingma and Ba,
tributed to the screenplay scenes that follow. In-
2014) for optimizing our networks. After experi-
tuitively, initial scenes of an episode tend to have
mentation, we chose an LSTM with 64 neurons for
high similarity with all other scenes in the screen-
encoding the scenes in the screenplay and another
play, and on their own are not very informative
identical one for contextualizing them. For the
(e.g., the crime, victim, and suspects are intro-
context interaction layer, the window l for comput-
duced but the perpetrator is not yet known). As
ing the surrounding context of a screenplay scene
a result, the undirected version of T EXT R ANK
was set to 20% of the screenplay length as pro-
tends to favor the first part of the story and the re-
posed in Papalampidi et al. (2019). Finally, we
sulting summary consists mainly of initial scenes.
also added a dropout of 0.2. For developing our
By adding extra importance to later scenes, we
models we used PyTorch (Paszke et al., 2017).
also encourage the selection of later events that
might be surprising (and hence have lower simi-
D Additional Results
larity with other scenes) but more informative for
We illustrate in Figure 5 the performance (F1 the summary. Moreover, in S UMMER, where the
score) of the directed neural T EXT R ANK and weights change in a systematic manner based on
S UMMER models in the unsupervised setting with narrative structure, we also observe that scenes ap-
respect to different λ1 values. Higher λ1 values pearing later in the screenplay are selected more
correspond to higher importance for the succeed- often for inclusion in the summary.
ing scenes and respectively lower importance for As described in detail in Section 3.3, we also
1932infer the narrative structure of CSI episodes in the
supervised version of S UMMER via latent TP rep-
resentations. During experimentation (see Sec-
tion 5), we found that these TPs are highly corre-
lated with different aspects of a CSI summary. In
Figure 6 we visualize examples of identified TPs
on CSI episodes during test time alongside with
gold-standard aspect-based summary annotations.
Based on the examples, we empirically observe
that different TPs tend to capture different types
of information helpful for summarizing crime in-
vestigation stories (e.g., crime scene, victim, per-
petrator, motive).
1933You can also read