Novel Computational Linguistic Measures, Dialogue System and the Development of SOPHIE: Standardized Online Patient for Healthcare Interaction ...
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
IEEE TRANSACTION ON AFFECTIVE COMPUTING AUGUST 2020 1 Novel Computational Linguistic Measures, Dialogue System and the Development of SOPHIE: Standardized Online Patient for Healthcare Interaction Education Mohammad Rafayet Ali, Taylan Sen, Benjamin Kane, Shagun Bose, Thomas M Carroll, Ronald Epstein, Lenhart Schubert, Ehsan Hoque, Member, IEEE Abstract—In this paper, we describe the iterative participatory design of SOPHIE, an online virtual patient for feedback-based practice of sensitive patient-physician conversations, and discuss an initial qualitative evaluation of the system by professional end users. The design of SOPHIE was motivated from a computational linguistic analysis of the transcripts of 383 patient-physician conversations from an essential office visit of late stage cancer patients with their oncologists. We developed methods for the automatic detection of two behavioral paradigms, lecturing and positive language usage patterns (sentiment trajectory of conversation), that are shown to be significantly associated with patient prognosis understanding. These automated metrics associated with effective communication were incorporated into SOPHIE, and a pilot user study identified that SOPHIE was favorably reviewed by a user group of practicing physicians. Index Terms—Virtual agent, Sentiment, Physician education, Communication skills, Physician-patient relations, Oncology, Cancer, Palliative care F 1 I NTRODUCTION E ffective patient-physician communication is fundamen- tal to a patient’s right to be fully informed and actively involved in health decision making. Good communication but is also rapidly customizable, cost-effective, and ready to be deployed online across geographical boundaries. The 2020 pandemic saw a dramatic increase in online skills facilitate physicians’ understanding of patients’ symp- interactions. In-person communications were aggressively toms, concerns, and treatment wishes [1]. Effective commu- replaced with virtual interactions across a wide spectrum nication in the clinical setting has further been correlated of domains from education to healthcare. Even the most with better patient health outcomes [2], [3], [4], [5]. Alterna- technologically inexperienced and averse, from preschool tively, a lack of effective communication has been associated students to senior citizens, were forced to learn and find with patients underestimating their disease severity [6] and the means to participate online. The medical education overestimating their prognosis [7]. Together, these findings system also felt pressure to accelerate physician training. suggest that training the physicians on the fundamentals of The loss of in-person interaction, may be likely to exacerbate how to communicate with patients, including taking turns, the current deficiencies in patient-physician communication. asking questions, showing empathy, and being positive is This problem is further complicated by the ever decreasing a very important part of medical education. In addition to amount of time physicians have to spend with their pa- in-person training, the state of the art medical education tients. Additionally, with increasing medical technologies to involves using trained actors to play the role of standard- learn and ever more specialized fields of medical training, ized patients who provide medical students feedback [8], physicians have less and less time for training in patient- [9], [10], [11], [12], [13], [14]. These techniques have the physician communication. limitations of being expensive in terms of time and money In this paper, we focus on patient-physician communi- as well as being prone to individual variation. Medical cation in cancer care. Communication between oncologists schools in the developing world may not even have the and patients is especially important due to the complexity resources to provide such training [15], [16]. There exists and the emotions involved in discussing the patient’s life a dire need to improve patient-physician communication expectancy. In addition, the oncologists need to explain training that is not only evidence-based and standardized, the severity of cancer, the multiple treatment options avail- able, and the correlates of patient involvement in complex • Mohammad Rafayet Ali, Taylan Sen, Benjamin Kane, Shagun Bose, decision-making [17], while expressing appropriate emotion Lenhart Schubert, and Ehsan Hoque are with the Department of and empathy. Despite decades of communication training Computer Science, University of Rochester, New York. and research, studies have shown that over 60% of late stage • Thomas M Carroll and Ronald Epstein are with the Department of Family cancer patients do not understand their prognosis [12]. It Medicine, University of Rochester Medical Center. is thus clear that identifying modifiable correlates of effec- Manuscript received August 2020. tive communication is essential for developing physician
IEEE TRANSACTION ON AFFECTIVE COMPUTING AUGUST 2020 2 communication training. We present a multi-stage research project leading to the development of an online virtual patient for physicians’ communication training. We begin with the development of automatic detection methods of two behavioral paradigms, lecturing and positive language usage patterns (the sentiment trajectory of conversation). We have used a data set that consists of 382 transcripts of con- versations between late stage (stage 3 or 4) cancer patients (Male=172, Female=210) and their physicians (Male=25, Fe- male=13) and a measure of each patient’s prognosis under- standing [7]. Computational linguistic analysis of the con- versation transcripts enabled us to develop automatic met- rics for evaluating the degree of lecturing-like structure in a conversation. In addition, we identify that most physicians tend to use one of three styles of varying their sentiment over time (i.e., there are three styles of sentiment trajectory). Fig. 1: A physician practicing communication skills with We show that these metrics have a significant association SOPHIE virtual patient. with patients’ level of prognosis understanding. We then developed an online virtual agent-based communication skills development system, SOPHIE (Standardized Online better prognosis understanding. Patient for Healthcare Interaction Education), which gives • Presentation of an iterative participatory design pro- users feedback on lecturing and positive language usage. cess, and an initial end-user evaluation with eight SOPHIE also provides feedback on the user’s speech rate practicing physicians, of an online virtual patient and number of questions asked. SOPHIE presents herself as (i.e., SOPHIE) for feedback-based practice of critical a late stage cancer patient and designed using the physician patient-physician conversations. communication training protocol – SPIKES [18]. Fig. 1 shows In this paper, in collaboration with oncologists and med- a physician practicing communication skills with SOPHIE ical educators from University of Rochester Medical Center in his home. SOPHIE allows users to practice in their own (URMC), we provide early ideas on how inspiration from af- private environment. fective computing and online interactions could potentially Prior interventions with physicians and patients have transform current medical education. promoted discussions about prognosis but have not im- proved prognostic understanding. We undertook the first set of analyses to discover patterns of communication that 2 R ELATED W ORK had not been previously described that might affect the This work encompasses on several interconnected areas outcome of prognosis conversations, with the intention of including, affect and sentiment analysis, prognosis un- applying findings from those analyses into the design of derstanding, patient-physician communication, virtual pa- SOPHIE. Since the target outcome was to improve the tients, and communication skills development programs prognosis understanding, the feedback was designed in [19]. Here we highlight the related research in these inter- such a way that has a direct association with the prognosis secting domains. understanding. This is why we developed SOPHIE utilizing Affective computing and sentiment analysis has been an existing data set. We first identify the affective compo- used by researchers for health-care monitoring and disease nents in the dataset on which we can give feedback such symptoms detection. Zucco et al. [20] proposed sentiment as sentiment and lecturing style of communication. We then and affective computing based architecture for depression validate it with statistical analysis. Finally, we implement detection. In subsequent work the authors [21] analyzed the feedback of SOPHIE using the knowledge we have from sentiment to detect the dropout of patients in tele-homecare our analysis. service. The association of positive patient outcomes with Our contributions include: physician affect has received limited experimental examina- tion and differing conclusions have been reached in regard • The development of an automated metric for mea- to the patient health outcomes. Hall et al. [22] found that suring the lecturing-like structure of a patient- the negative affects of the physicians such as showing anger physician conversation transcript, and anxiety are correlated with patients’ contentment. In • The identification that most doctors use one of three contrast, Verheul et al. [23] in a study with 30 female patients styles of sentiment trajectory (i.e., pattern of modify- found that warm and empathetic communication helped ing their positive language usage over the course of decrease the state of anxiety among patients. Similarly, Di a patient-physician conversation). Blasi, et al. [24] found in their review of 25 randomized con- • Demonstration that the degree of lecturing structure trolled trials on affective physician communication, show is significantly associated with the level of prognosis inconsistency regarding emotional and cognitive care. Sen et misunderstanding. al. [25] also found a lack of association of overall conversa- • Finding that a certain sentiment trajectory style (one tional positive sentiment with patient ratings of their oncol- which involves delivering technical information and ogists’ communication skills. Prior studies have also studied ending with positive language) is associated with the association of physician affect on patient information
IEEE TRANSACTION ON AFFECTIVE COMPUTING AUGUST 2020 3 recall, prognosis understanding, and better health outcomes TABLE 1: Study Data: Counts and Prognosis Survey Options in general. In a study of women receiving simulated breast Resp. # Description cancer-related communications from a videotaped oncolo- 0 100% gist, van Osch et al. [26] found that affective communication 1 about 90% improves information recall. When physicians used positive 2 about 75% 3 about 50-50 affect statements participants provided significantly more 4 about 25% correct answers on a questionnaire testing the participants’ 5 about 10% recall of details in the diagnosis, prognosis, and treatment 6 0% X don’t know options. A similar study involving participant viewing of videotaped simulated oncologist communications, Shapiro refine the program’s feedback module. et al. [27] found that participants who received communica- tion from a worried physician as opposed to the standard, recalled significantly less medical information. The use of 3 M ATERIALS negative and positive affect does not seem to have a consis- We performed a post-hoc analysis of a study ( [38]) involv- tent effect on patient-physician communication. This is why ing 382 visits between cancer patients (N = 382) and their we should focus on not only the overall affective state but oncologists (N = 38). The data includes a transcript of the also the timing of the affect. The importance of sentiment conversations, in addition to both patient and physician variation over time has long been recognized in storytelling surveys associated with each visit. The survey included [28] and more recently has been shown to be relevant in questions to the physician and to the patient regarding the natural language analysis and face-to-face communication patient’s prognosis [39]. Specifically, the prognosis question [29], [30], [31]. directed to the physicians was: ”What do you believe are the In health care communication skills training, virtual chances that this patient will live for 2 years or more?”; the agents, and online platforms have been used in an at- options provided for a response are shown in Table 1. tempt to provide an effective and reproducible experience. Patients were separately asked “What do you believe your In the past, affective computing helped design intelligent doctor thinks are the chances that you will live for 2 years virtual agent-based interactions for tele-health [32], [33]. or more?”, with the same options for a response. When Prendinger et al. [34] presented their initial work on a the absolute difference of the responses is greater than 1, virtual character that analyzes physiological data in real- the patient-physician prognostic conversation is defined as time, interprets emotions, and addresses users’ negative being misunderstood. Data in which either the physician or affective states with empathic feedback. Peddle et al. [35] de- patient refused to answer were not used. veloped a virtual patient (VP) to develop and practice non- technical knowledge, skills, and attitudes among under- graduate health professionals. In a study with second and 4 M ETHODS third-year nursing students, the authors found that interac- In patient-physician communication there are several be- tions with VPs developed knowledge and skills across all havioral paradigms that help prognosis understanding. categories of non-technical skills to varying degrees. Third- Among many behavioral paradigms, we have explored two year students suggested that interactions with VPs helped patterns of behavior – lecturing, and the sentiment trajectory develop knowledge and skills in a clinical setting. Angus of conversation. We first present how we set about detect- et al. [36] developed a graphical visualization tool to model ing these phenomena automatically and determining how patient-physician dialogue, to identify patterns of engage- they are associated with prognosis understanding. Then ment between individuals including communication accom- we explain our feedback design for these two behavioral modation, engagement, and repetition. Kleinsmith et al. [37] patterns, applicable in conversation practice with a virtual developed a chat-based interactive virtual patient for early- conversational agent. stage medical students to practice empathetic conversation. During the training, students can gather information regard- ing the history of the present illness, medical history, family 4.1 Lecturing history and social history. Additionally, during each session, Lecturing generally occurs when the physician delivers a the VPs delivered a statement of concern. These statements, lot of information without giving the patient a chance to termed empathetic opportunities, were designed to elicit ask questions or to respond [1], [40]. We developed an an empathetic response from the user. In a study, medical algorithm for calculating the LECT-UR Score (Lecturing students interacted with the VP and standardized patients. Estimation through Counting Turns with an Unbalanced- The responses of the participants were then rated by coders, length Ratio), a measure of lecturing-related conversational and it turned out that responses were more empathetic with structure. The algorithm compares the number of words virtual patients than with standardized patients. spoken by the physician to the number of words spoken In this work, we have focused on improving prognosis by the patient across a sliding window of a number of understanding among late-stage cancer patients. To this patient-physician turns. When the average number of words end, we designed a virtual patient to conduct conversations spoken by the physician exceeds a given threshold, while with oncologists. To provide feedback to users on com- the average number of words spoken by the patient is munication skills, we first developed algorithms to detect below the threshold, the conversation segment is counted behavioral cues in patient-physician conversations and then as a lecturing event. Fig. 2 shows the area where a lecturing engaged practicing physicians in participatory design to event can occur in the space of the number of words spoken
IEEE TRANSACTION ON AFFECTIVE COMPUTING AUGUST 2020 4 Fig. 2: Classification of Physician (D) and Patient (P) Turn Lengths over Window as Lecturing and Not-Lecturing a) Regions of Lecturing and Not-Lecturing as P length vs. D length b) example transcript turn lengths by the physician (y-axis) and patient (x-axis). The thresholds approaches zero. Similarly, as the probability of an event are determined by maximizing the entropy of the outcome xi approaches zero, the contribution of such events to the variables (i.e., prognosis misunderstanding). total information content in the data approaches zero. Thus, As shown in Fig. 2 region 1, when both the physician in order to maximize the information contained in the and patient speak with brief turns it is not counted as LECT-UR score, the scores should be well distributed (i.e. an instance of lecturing. Similarly, in region 3, when the maximizing the entropy). patient is speaking with a long turn length is not labeled In order to determine the optimal τ and W , we perform as lecturing. Only when the physician’s average turn length a grid search. For a given τ and W we first calculate the exceeds a threshold, and the patient’s average turn length LECT-UR score L based on equation 1. We then applied does not exceed the threshold, (i.e., Region 2), is the window the kernel density estimation method [42] to compute the labeled as an instance of lecturing. probability density function P (x). From the probability This algorithm is expressed in the following equations: density function we then obtain the entropy of L using k+W ! k+W ! equation 2. In Fig.3, the entropy values for different values X X X of τ and W are shown. The maximal entropy occurs with L= I ωi − τ × I τ − ωi (1) τ = 103 and W = 20. After calculating the LECT-UR score ∀k i=kwi ∈D i=kwi ∈P with the optimal parameters for each office visit transcript, we partition the data into high and low LECT-UR groups I(x) = { 0 : x < 01 : x ≥ 0 based on the median value. We then use the Z-score two- tailed population proportion test to see the difference in the where, percentage of prognosis misunderstanding. L : LECT-UR Score To understand the effects of the confounding variables W : window length in number of turns we performed a logistic regression analysis. We applied τ : turn length disparity threshold logistic regression on gender, age, disease severity, average ω : words in the transcript sentiment of the conversation, study site, study arm, and D : physician utterances the LECT-UR to predict the percentage of prognosis misun- P : Patient utterances derstanding. We first normalized the independent variables and fit a logistic regression model predicting the prognosis Referring to equation 1, a value for the τ parameter must understanding. In the section 5 we present the regression be determined. As τ approaches zero, the area of region 1 weights and the expected prognosis misunderstanding per- in Fig. 2 will also approach zero. Alternatively, if a very centage for different quantiles of the LECT-UR score. large value is used for τ , region 1 will cover the entire data space. In order to be useful, the LECT-UR score should have 4.2 Sentiment Trajectory variability. Borrowing concepts from information theory, the amount of information in a signal can be measured by the The sentiment of a text segment, generally represents the signal’s entropy, where entropy is a measure of the amount emotional tone of the segment. In this work, we focus on of uncertainty [41]. More specifically, for a given data set X , positive language usage. We define the sentiment trajectory the definition of the entropy, H(X), is: as the change that occurs in physician positive sentiment over the course of the conversation. Prior research suggests n X 1 that the change of affective states is more important than H(X) = P(xi ) logb (2) the overall affective state. For example, Ali et al. [30] showed i=1 P(xi ) being positive at the beginning and at the end of a conversa- where P (xi ) represents the probability of observing tion is more effective than being positive overall. However, the it h data point. As the probability of an event xi ap- the physician sentiment trajectory over a conversation has proaches certainty (i.e. P (xi ) ≈ 1), the information content not been well-studied in the context of patients’ prognosis
IEEE TRANSACTION ON AFFECTIVE COMPUTING AUGUST 2020 5 TABLE 2: Average Prognosis Misunderstanding scores in High and Low LECT-UR Groups Prognosis Group p-value effect size Misunderstanding % High LECT-UR 83.6 0.00058 0.37 Low LECT-UR 72.3 we applied the inference test for population proportions pairwise between the groups. We analyzed the effects of confounding variables by performing a logistic regression analysis with the sentiment trajectory styles. Specifically, we applied logistic regression on gender, age, disease severity, average sentiment of the conversation, study site, study arm, and the conversation Fig. 3: Finding the Optimal Lecturing Threshold and win- styles to predict the outcome measures. After fitting data dow size based on the heatmap of Entropy. to logistic regression, we again can compare the relative effect that each of the input variables has on predicting understanding. First, we describe how we define sentiment whether a given data point (conversation) results in a “Don’t trajectories and identify the number of sentiment trajectory understand prognosis” classification. After normalizing the styles. Later, we present the association between the trajec- inputs (i.e., scaling and shifting to have mean=0 and vari- tory styles and prognosis understanding. ance=1) we fit the model and hence find the model weights. We then investigate the weights of the logistic models and We utilized the VADER ( [43]) automatic text analysis the prognosis misunderstanding percentage for each of the tool. VADER calculates sentiment through the use of a rule- conversation styles. based model that employs a sentiment lexicon (dictionary In addition to the binary prognosis misunderstanding of words containing an associated valence measure). The we have looked at the linear score of misunderstanding. We VADER positive sentiment feature is the result of a large performed a linear regression analysis. The details are in number of human raters’ understanding of positive and Appendix A and B. negative emotion associated with particular words. The VADER positive sentiment score was evaluated for each turn of the conversation. These physician and patient senti- 5 F INDINGS ment scores were used in two ways —- 1) average analysis, 5.1 Association between LECT-UR Score and Progno- and 2) sentiment trajectory. The average sentiment scores sis Understanding for the physician were calculated for each transcript. The As shown in Table 2, the High LECT-UR Score group has transcripts were split into two groups based on the median a larger percentage of prognosis misunderstanding than the of the physician average sentiments (i.e. a High Sentiment Low LECT-UR Score group (83.6 vs. 72.3) with a correspond- group and a Low Sentiment group). The outcome mea- ing p-value of 0.00058 and an estimated Cliff’s d effect size sure (Prognosis Misunderstanding%) was then compared of 0.37 [46]. between the two groups using the z-score population pro- Fig.5 shows the logistic regression weights when pre- portion test. dicting the prognosis misunderstanding %. The (*) marked The sentiment trajectory is defined as the time series of features had a p-value less than 0.05. Among all the features average physician positive sentiment over the segmented the disease severity had the highest positive correlation with conversation. More specifically, we partitioned each con- the prognosis misunderstanding. This shows that the more versation transcript into a number of non-overlapping seg- the disease has progressed the more the patients are likely ments and calculated the physician’s average positive sen- to misunderstand their prognosis. Although the LECT-UR timent within each segment. Each conversation’s sentiment score has small positive weight than age and severity, it trajectory is represented as a multidimensional vector, each was significant. This model thus suggests that the higher a dimension corresponding to the average sentiment within conversation’s LECT-UR score, the more likely a patient will a corresponding segment of the conversation. We next misunderstand their prognosis. Fig. 6 shows the prognosis determined whether distinct styles of physician sentiment misunderstanding percentage for the different quantile val- trajectory existed among the conversations and investigated ues of the LECT-UR score. To understand this let’s select whether any of these physician styles demonstrated signif- a quantile value of LECT-UR. For example, the oncologists icant differences in prognosis understanding. To determine who are above the 80th percentile based on their LECT-UR whether distinct styles of sentiment trajectory exist, we ap- score had more than 54% of patients fail to understand their plied the k-means clustering algorithm ( [44]). The number prognosis. of clusters k was selected using the widely used Silhouette method ( [45]), in which a grid search over a finite space of integer values for the k parameters is performed to maxi- 5.2 Association between Sentiment and Prognosis Un- mizes the Silhouette score. Fig. 4 shows the steps of finding derstanding the sentiment trajectories. In order to determine whether The difference in the prognosis misunderstanding % be- any of the resulting sentiment trajectory clusters had sta- tween the high and low average positive sentiment groups tistically significant differences in the outcome measures, did not show a significant difference. Out of the analyzed
IEEE TRANSACTION ON AFFECTIVE COMPUTING AUGUST 2020 6 Fig. 4: Sentiment Trajectory Analysis Steps. (From left to right) 382 physician-patient conversations transcribed, VADER tool is used to calculate the positive sentiment of each physician turn, the full conversation is segmented into equal regions, the physician sentiment in each region is averaged into a trajectory, 382 trajectories are clustered in three clusters (i.e. trajectory styles) that best fit the data, statistical comparisons was done of the patient prognosis understanding in each cluster. Fig. 7: Resulting Sentiment Trajectory Clusters for best K=3. TABLE 3: Prognosis misunderstanding % in three sentiment Fig. 5: Logit model weights for predicting whether the trajectory clusters. prognosis is misunderstood. Trajectory cluster (size) Pairwise statistical comparison A (15) B (58) C (191) PAB PBC PAC 46.1 52.6 67.9 0.34 0.04 0.06 a more dynamic shape, with increases in positive sentiment at 25% into the conversation (segment 2), as well as at the end of the conversation (segment 7). By contrast, Clusters B (Medium) and C (low) have a mostly flat sentiment level throughout the conversation with approximate aver- age VADER sentiment levels of 0.1 and 0.05 respectively. Shown in Table 3 are the outcome measures for each of the three trajectory cluster groups along with pairwise population percentage inference test p-values. As shown by the Prognosis Misunderstanding %, the low cluster (cluster C) showed the highest percentage with 67.9 % of the patients Fig. 6: Prognosis misunderstanding percentage for the dif- having a discordant understanding of their prognosis. The ferent quantile values of the LECT-UR score. p-values for comparing the percentages between low and dynamic and low with medium clusters were 0.04 and 0.06 number of clusters (k = 2 through 10), the number of respectively. trajectory clusters that had the highest Silhouette score was Fig. 8 shows the logistic regression weights when pre- k=3. In addition, the BIC (Bayesian information criterion dicting the Prognosis Misunderstanding %. The variables [47]) analysis also identified that the optimal value for k is marked with a (*) had p < 0.05. In Fig. 8 the highest 3. Shown in Fig. 7 are the resulting three trajectory clusters: positive value was assigned to severity. Although this is not cluster A (red, n = 15); cluster B (orange, n = 58), and cluster significant, it indicates that patients with a higher severity C (blue, n = 191). Cluster A (Dynamic) is characterized by level of the disease are more likely to misunderstand their
IEEE TRANSACTION ON AFFECTIVE COMPUTING AUGUST 2020 7 ronment, which would be difficult to achieve with actual standardized patients. 6.1 Scenario We have developed a prototype of the SOPHIE program, which allows individuals to have a conversation with a virtual agent concerning prognosis and treatment options. SOPHIE presents herself as a late-stage cancer patient. We have used the SPIKES protocol to guide the conversation [18]. The SPIKES protocol was developed to train physicians deliver bad news. The SPIKES protocol has six steps – 1) setting up the interview, 2) assessing patients’ perception, 3) obtaining patients’ invitation, 4) giving knowledge and in- Fig. 8: Logistic regression model weights for predicting formation to the patient, 5) addressing the patient’s emotion whether the prognosis is misunderstood. with empathetic responses, and 6) strategy and summary. With SOPHIE, at the beginning of the conversation (SPIKES TABLE 4: Confounder-Adjusted Logit Model step 1) SOPHIE introduces herself and mentions that she Trajectory Cluster PMU % β p-val has lung cancer. Then SOPHIE raises the topic of her sleep A (Dynamic) 49.76 -0.294 0.033 pattern at night and asks if she needs to change her pain B (Medium) 70.74 -0.155 - medication, allowing the physician to assess her perception C (Low) 84.85 0.209 - (SPIKES step 2). She states that her current pain medica- tion, Lortab, is not working anymore. SOPHIE then turns prognosis. Patient gender had negative weight which in- attention to her test results, giving the physician a chance dicates that female patients were more likely to misunder- to obtain SOPHIE’s invitation to talk about more difficult stand their prognosis. This is also true for physician gender topics (SPIKES step 3), before asking more specifically about but not significant. Average physician sentiment has low her prognosis if the physician did not already address it, positive weight but significant. This indicates being positive thus allowing the physician to provide information to the overall is associated with misunderstanding prognosis. This patient (SPIKES step 4). SOPHIE then asks about what finding is similar to what we have seen in the past where her options are, allowing the physician to give empathetic being positive had a negative correlation with how the responses (SPIKES step 5). Finally, she follows up by dis- patients rate their physicians [25]. Among all the clusters, cussing whether chemotherapy remains an option, whether the dynamic cluster has the lowest value. This indicates she should focus on comfort care, what the side effects of that when physicians used the dynamic sentiment pattern chemotherapy are, and how to break the news to her family, throughout the conversation, the patients were less likely to allowing for the physician to provide strategy & summary misunderstand their prognosis. information (SPIKES step 6). Unlike linear regression, with logistic regression there This type of discussion promotes understanding of the is no simple way to adjust the output (i.e., “correct” the patient, gathering information from the patient, discussing output) for the effect of confounding variables of each data critical information, and responding with empathy. point. This is because the actual outputs are binary, whereas the model output is a probability. Instead, we can compare the predicted model Prognosis Misunderstanding % for 6.2 Dialogue System each cluster. When all confounding variables are set to have The SOPHIE program is built on top of Eta, a general the average value over our data set, we compute the models’ purpose dialogue management framework representing a predicted Prognosis Misunderstanding % for each cluster further development of the LISSA system [48], [49], [50]. (see table 4). The Wald test p-value of the logistic regression Each dialogue agent built within the Eta framework defines is also shown in table 4 (marked * in Fig. 8 when p < 0.05). a flexible, modifiable dialogue schema, which specifies a This again indicates that with confounding adjustment, the sequence of intended and expected interactions with the dynamic style cluster has the lowest Prognosis Misunder- user. The body of a dialogue schema consists of a sequence standing % among all clusters. of formal assertions that express either actions intended by the agent, or inputs expected from the user. These events are dynamically instantiated into a dialogue plan 6 D ESIGN OF SOPHIE over the course of the conversations. As the conversation Our aim is to develop a virtual standardized patient for proceeds, this plan is subject to modification based on the practicing communication skills. In medical education, stu- interpretation of each user input in the context of the agent’s dents practice with a standardized patient – an actor/actress previous utterance. For instance, if a planned query to the pretends to have a medical condition. Students interact with user has already been answered by some part of a user’s the standardized patients and later they receive feedback on previous input, the dialogue manager can skip that query. their interaction. Our goal is to allow the medical students The dialogue manager can also expand steps into subplans to practice their communication skills with a virtual agent, by instantiating sub-schemas in the case of more complex allowing multiple repetitions in each student’s own envi- interactions.
IEEE TRANSACTION ON AFFECTIVE COMPUTING AUGUST 2020 8 The dialogue management framework captures the users’ response from the audio stream using an automatic speech recognition technique. Both interpretation of the user’s replies and generation of the agent’s responses are handled using transduction to and from simple context- independent English sentences called gist-clauses. The dia- logue manager interprets each user’s input in the context of SOPHIE’s previous question, using this context to select topically relevant pattern transduction hierarchies to use to interpret the user’s response. The context of the previous question is useful for resolving anaphora, ellipsis, and other pragmatic phenomena. The rules in the selected hierarchies are then used to derive one or more gist-clauses from the user’s input, containing explicit representations of both statements and questions detected in the user’s utterance. For example, if SOPHIE asks “Do you think I should take Fig. 9: SOPHIE virtual agent stronger pain medication?” and the user answers “Yes.”, the 6.3 Interface gist-clause extracted would be “I think you should take stronger pain medication.” If the user replies “Can you The SOPHIE system features a virtual agent (shown in fig. tell me more about how you’re feeling?”, the gist-clause 9). At the beginning, users start the conversation by pressing extracted would be “Can you tell me more about your the “start recording” button. Users can then proceed to pain?”, having interpreted the question as an inquiry about conversing with SOPHIE, and when the conversation is SOPHIE’s pain in the particular context of her question. over the program takes the user to the feedback page. The feedback interface is shown in Fig. 10. On the left side As mentioned, the gist-clauses are derived using hierar- of the feedback interface we show the transcript. The red chical pattern transduction methods. Each transduction hi- marked speech is considered too long for the patients. On erarchy specifies patterns at its nodes that are to be matched the right side of the interface we show the speech rate of to input, with terminal nodes providing result templates to the user, number of questions user asked, turn taking, and be used according to various directives (e.g. storing as a the sentiment trajectory. Past literature has established that gist-clause, outputting the result, specifying a sub-schema conversational speech rate is important in enabling patients to be instantiated, etc.). The pattern templates look for to understand their prognosis. Also, asking questions of the particular words or word features, including “wildcards” patient is important for ensuring that the patient under- matching any word sequence of some length. In the case of a stands what is being said ( [1]). The turn taking annotation failure to match, the system first tries siblings of the pattern shows the length of each turn by SOPHIE and the user. before backtracking to the previous level; the efficiency of The example was chosen to illustrate the lecturing style of the hierarchical pattern matching approach lies in the fact conversation. The feedback shows the sentiment trajectory that higher levels can segment utterances into meaningful of both SOPHIE and the user. Additionally, the feedback parts, thus reducing the amount of backtracking necessary shows a suggested sentiment trajectory for the user. The to interpret the user’s input. feedback page displays explanations of individual items The agent’s responses to the user are likewise deter- when users hover their mouse on them. mined using hierarchical pattern transduction. In the case where the gist-clause from the user’s utterance is a simple 6.4 Pilot Study statement, the agent selects a reaction to the gist-clause and We conducted a pilot study with nine practicing clinicians either instantiates a sub-schema to ask a follow-up question, (fellows, residents, and nurse practitioners) from the Uni- or proceeds to the next topic in the main schema. If the gist- versity of Rochester Medical Center. Participation was vol- clause from the user’s utterance is a question, the agent untary and we did not offer any payment for their participa- instantiates a sub-schema to select a reply to the user’s tion. Additionally, we made it clear that not participating or question and await either a follow-up question or closure stopping the study in the middle will have no consequences. from the user. The system also has the potential to form Among these participants, one participant dropped out due replies to multiple gist clauses from a single user turn, for to the bad audio quality of her computer. All participants instance reacting to the user’s statement before responding were white and aged between 30 and 55. Three participants to a final question by the user. were female and all were native English speakers. Our goal The transduction hierarchies themselves were designed was to gather more information about their experience with in a modular fashion, with a “backbone” of transduction SOPHIE, any limitations, and how we could improve the trees detecting general questions that SOPHIE might expect system. The study was performed with one participant at a a user to ask, with additional transduction trees for detect- time on the Zoom communication platform. Each day, we ing questions and replies specific to the current topic of the asked the invited participant to have a conversation with conversation. In the case of a failure to match a specific SOPHIE and to look at the feedback. response, the dialogue manager can fall back to the current After conversing and receiving the feedback, the par- general question, and if this fallback fails, simply output a ticipants were interviewed. We have performed a thematic generic default response. analysis on the interview transcripts; our findings are below.
IEEE TRANSACTION ON AFFECTIVE COMPUTING AUGUST 2020 9 Fig. 10: Feedback interface of SOPHIE. On the left side the conversation transcript is shown. On the right (from top to bottom) speech rate, number of questions are shown. Turn taking shows the turn length and at the bottom the sentiment trajectory of both physician and SOPHIE are shown with the ideal/suggested sentiment trajectory. 6.4.1 Medical History questions did not really fit and she just jumped to All the participants mentioned that a brief medical history the next topic so I jumped with it but that was a should be presented before starting the conversation with little bit jarring to me.” SOPHIE. One participant said, Although SOPHIE changed the conversation topics “I think some kind of medical record would be abruptly, the questions she asked were found to be realistic. extremely helpful. I thought I don’t have any in- Five participants felt that SOPHIE was able to express formation to say to her.” her concerns and make them feel the seriousness of the The participants mentioned that in a regular standardized situation. patient visit, they are given a medical record before they go into the room. They suggested the same scenario should 6.4.3 Feedback on Speech Rate be replicated for SOPHIE. They also mentioned that the way SOPHIE initiated presentation of her symptoms was Participants (seven out of eight) mentioned that the speech unusual. In most cases, patients do not actively start the rate feedback was easy to understand and very useful. One conversation. Rather, the physician looks at the patient’s participant said, medical record and then starts asking about any new symp- “I know I tend to speak very fast, receiving feed- toms. In the future, we expect to modify the dialogues back on my speech rate is going to be very useful.” so that SOPHIE appears more passive and lets the users ask questions, though completely user-driven conversation Another participant mentioned that in normal practice there remains beyond the state of the art. is no way of measuring the speech-rate. However, with SOPHIE we could provide the information about how fast 6.4.2 Topics of Conversation the physicians are speaking, which is useful. Participants (four out of eight) mentioned that SOPHIE “I think the feedback (speech-rate) was useful. I jumped between topics and did not allow full coverage of a never had someone measure my speech rate before. given topic. For example, SOPHIE begins talking about her Sometimes I try to be cognizant of speaking a pain medication, but the participants often asked questions little bit slower with the patients but it was nice about the current dosage and about other pain medication to actually get some feedback like you are doing she had taken in the past. Since SOPHIE’s limited dialogue okay.” repertoire falls short of covering those questions, she starts However, the participants also noted that SOPHIE’s talking about her current medication (i.e., Lortab). One speech rate was constant, making it difficult for them to participant said, adjust their speech rate depending on whether they are “When she mentioned pain and I was trying to find discussing serious issues or a casual topic. In the future, more about the pain in order to help her with her we plan to adjust SOPHIE’s speech rate based on the seri- question. But the answers that I gave her to her ousness of the topic being discussed.
IEEE TRANSACTION ON AFFECTIVE COMPUTING AUGUST 2020 10 6.4.4 Number of Questions Asked the development of SOPHIE and the use of such a system Seven out of eight participants expressed that feedback as a basis for feedback. on the number of questions asked was very useful. One Our finding of associations between trajectory styles participant said, and lecturing tendencies with prognosis understanding “It was helpful to get the information about how measures may not be causal. The lecturing analysis was many questions you have asked, because I think motivated from prior research that suggested that when a a lot of the times we walk away from the conver- physician tends towards lecturing, this results in the patient sation thinking that we really invited the patients not retaining as much of the information presented [1]. into the talk, when maybe we didn’t and we did a An alternative explanation could be that when physicians lot of lecturing.” sense that patients do not understand, physicians are mo- tivated to speak more, explain in greater detail, leading to In addition to the number of questions asked, participants a more lecturing-like structured conversation. Additionally, suggested that it should be highlighted what type of ques- apparently passive patients may just lack understanding, tions were asked, for example, how many history-taking which can result in poor engagement, and this may result in questions were asked and how many emotional questions conversations with a high LECT-UR score. Additionally, the were asked. extent to which the LECT-UR score correlates with human annotated instances of “ground truth” lecturing should be 6.4.5 Explanation of Sentiment investigated. However, it should be noted that despite any Seven participants mentioned that they did not understand difference between the LECT-UR lecturing-like structure the meaning of the sentiment values. They also said that measure and human-labelled ground truth instances of lec- the sentiment feedback is hard to interpret and they often turing, our results establish that the LECT-UR score serves confused it with empathy. Four participants wanted to see as a useful metric in its association with patient prognosis an example sentence of positive and negative sentiment. misunderstanding. The participants also mentioned that changing or adjusting In explaining the association of higher prognosis un- sentiment while engaged in the conversation may add to derstanding with the dynamic sentiment trajectory style, the cognitive load. They suggested that instead of asking we surmise that being dynamic keeps the patient more the user to be positive at certain moments we should just engaged, and that ending on a positive note keeps the stress the importance of dynamically adjusting sentiment. patient less depressed and more likely to remember the information just presented. However, again, an alternative 6.4.6 Additional Feedback anticausal explanation could be that patients’ lack of progno- The participants also asked for additional feedback that they sis understanding, and their physician’s perception of this, found useful in practice. Two participants said that there motivates the physician to speak in a calmer, less dynamic are a few expressions of empathy in the dialogue and they way (e.g., sentiment trajectory styles B or C). should be highlighted in transcripts so that users could look We found that three clusters (k=3) represent the data back and understand how they responded. One participant best according to the Silhouette score. While the Silhouete said that the turn-taking feedback is useful, however, it does score is trusted method for finding the optimal number of not show the total amount of time a person was speaking. clusters for k¿=2, the Silhouette method is unable to evaluate The participant said, when the data is better represented by a single group (i.e. “I tend to speak a lot, but I don’t want to make the k=1). In order to determine that our finding of three clusters patients feel that I am not listening. I want to know is not an artifact of the techniques used, we used the that I am giving a chance to ask questions.” Bayesian Information Criterion (BIC) [47] as an additional He suggested adding a bar chart to the feedback page that method of validating the optimal k. We used a related indicates the total speech length and time. clustering technique, the Gaussian Mixture Model (GMM), Three participants suggested giving feedback on nonver- together with BIC to determine whether the data is better bal behaviors, such as eye contact. One of them said, represented by a single cluster. The GMM-BIC analysis also “One of the things I think is important, and I have found that the optimal k=3, and importantly showed that seen it in other clinicians, is eye contact. I think it’s k=1 was inferior. While it is possible to use a Gaussian super important when we are giving bad news or Mixture Model as our primary clustering method instead having difficult conversations.” of k-means, there are multiple reasons why k-means is more appropriate such as skewed distribution and fixed intervals. In addition to considering the number of clusters we have 7 D ISCUSSION , L IMITATIONS , AND F UTURE W ORK experimented with a range of values for the number of We have described two novel contributions to communi- segments. A large segment is not suitable for trajectory cation research; empiric associations between automatically analysis since it may contain the bulk of the conversation, detected behaviors and patient prognostic understanding, and a small segment size is also not suitable since it may and the development of SOPHIE, an automated system not contain representative turns from both physicians and for teaching and evaluating patient-physician communica- patients. Thus we experimented with five, eight, ten, and tion. In addition to the communication training program, fifteen as our number of segments. In this paper, we have the automatic detection of behaviors can be applied in shown results for the choice of eight segments, omitting the prerecorded standardized patient interactions to evaluate others as they produced similar results. communication skills. We acknowledge some limitations in Some limitations exist with regard to the bigger picture
IEEE TRANSACTION ON AFFECTIVE COMPUTING AUGUST 2020 11 of SOPHIE-like virtual agents. Past research suggests that boundaries. This will promote the fair use of the program by while conversing with a virtual agent or AI-driven conver- reaching the lower socio-economic areas. Indeed, we believe sational agent, humans tend to use shorter turns [51]. This that successful SOPHIE-like systems could have broader could be a limitation of using SOPHIE to train users to avoid global impacts. Two-thirds of cancer deaths happen in low- lecturing, since users might use shorter turns regardless of and mid-income countries such as those in Latin America feedback. Our LECT-UR scoring method utilizes a window and sub-Saharan Africa [15], [16]. However, most of the of consecutive turns that also includes the virtual agent’s seriously ill patients don’t have access to quality palliative turn. This allows the lecturing feedback to dynamically care (PC) because of inadequate PC training programs. adapt to the conversation states and to the user’s behavior. Current medical training in the countries of these regions We think that this can help circumvent the limitation posed focuses on treating diseases. Comfort care in chronic life- by using feedback trained on human-human conversation threatening diseases such as cancer is still in its infancy. with a computerized dialogue system, though addressing In Africa, some countries–Kenya, Uganda and Botswana– this concern through a randomized study remains part of have initiated post-graduate training programs for palliative our future work. care [54], [55]; only South Africa has a well-established post- The current dialogue manager itself also has some lim- graduate and research program on palliative care [56]. We itations, which we aim to address in the future. First, the are hopeful that online programs such as SOPHIE can pro- output of the currently used automatic speech recognition vide a basis for helping these communities develop training (ASR) software1 does not include punctuation. This limits programs for PC physicians. the agent’s ability to interpret the user correctly. Secondly, as discussed in Section 6.4.2, the dialogue manager tended to abruptly jump to the next topic in the main dialogue 9 ACKNOWLEDGMENTS schema in cases where it failed to understand the user’s input. This will be addressed by further expanding the This research was partially funded by the following grants interpretation patterns on the basis of the dialogues we NSF NRT-DESE Award 1449828 and in part by grant observed in this study, as well as by allowing for more W911NF-19-1-0029 with the US Defense Advanced Research robust default strategies, such as staying on topic when Projects Agency and the Army Research Office. it appears that the agent misinterpreted the user’s input or when the user’s input appears irrelevant to the agent’s question. R EFERENCES While our study focused on high patient prognosis un- [1] A. L. Back, R. M. Arnold, W. F. Baile, J. A. Tulsky, and K. Fryer- derstanding as a positive goal, it should be acknowledged Edwards, “Approaching difficult communication tasks in oncol- that patients sometimes don’t want to know specifically ogy 1,” CA: a cancer journal for clinicians, vol. 55, no. 3, pp. 164–177, 2005. how much time they have left [52]. In designing a commu- [2] S. H. Kaplan, S. Greenfield, and J. E. Ware Jr, “Assessing the nication training program we should incorporate options effects of physician-patient interactions on the outcomes of chronic as to how much information the physician should deliver. disease,” Medical care, pp. S110–S127, 1989. Another limitation of this work may be that our findings are [3] E. W. Nawar, R. W. Niska, and J. Xu, “National hospital ambula- tory medical care survey: 2005 emergency department summary,” limited to patient-physician relationships involving diseases 2007. and conditions as serious and sensitive as advanced cancer [4] J. Oates, W. W. Weston, and J. Jordan, “The impact of patient- care and end-of-life communication. centered care on outcomes,” Fam Pract, vol. 49, no. 9, pp. 796–804, 2000. [5] R. S. Beck, R. Daughtridge, and P. D. Sloane, “Physician-patient communication in the primary care office: a systematic review.” 8 C ONCLUSION The Journal of the American Board of Family Practice, vol. 15, no. 1, In summary, in this paper, we provide early results of our pp. 25–38, 2002. [6] J. C. Weeks, P. J. Catalano, A. Cronin, M. D. Finkelman, J. W. Mack, multi-stage research examining patient-physician conversa- N. L. Keating, and D. Schrag, “Patients’ expectations about effects tions, identification of effective traits (not lecturing, asking of chemotherapy for advanced cancer,” New England Journal of questions, delivering news on a positive note), development Medicine, vol. 367, no. 17, pp. 1616–1625, 2012. of an automated way of evaluating these traits, and the [7] R. Gramling, K. Fiscella, G. Xing, M. Hoerger, P. Duberstein, S. Plumb, S. Mohile, J. J. Fenton, D. J. Tancredi, R. L. Kravitz design of a real-time online standardized patient-physician et al., “Determinants of patient-oncologist prognostic discordance communication training system where an avatar plays the in advanced cancer,” JAMA oncology, vol. 2, no. 11, pp. 1421–1426, role of a standardized patient. We structured our exploration 2016. in the context of conversations between final stage cancer [8] A. Eid, M. Petty, L. Hutchins, and R. Thompson, ““breaking bad news”: standardized patient intervention improves communica- patients (i.e., terminal patients) and oncologists. tion skills for hematology-oncology fellows and advanced practice In [53] McGreevey et al. presented a few considerations nurses,” Journal of Cancer Education, vol. 24, no. 2, pp. 154–159, for implementing AI-driven conversational agents in health 2009. [9] A. K. Sachdeva, P. J. Wolfson, P. G. Blair, D. R. Gillum, E. J. Gracely, care. One important consideration is the level of risk associ- and M. Friedman, “Impact of a standardized patient intervention ated with a conversational agent when it makes a mistake. to teach breast and abdominal examination skills to third-year SOPHIE is a low-risk program, and can be augmented with medical students at two institutions,” The American journal of traditional training modules. In addition to being low-risk, surgery, vol. 173, no. 4, pp. 320–325, 1997. [10] M. Ju, A. T. Berman, and N. Vapiwala, “Standardized patient SOPHIE allows access by individuals beyond geographical training programs: an efficient solution to the call for quality im- provement in oncologist communication skills,” Journal of Cancer 1. https://www.nuance.com/index.html Education, vol. 30, no. 3, pp. 466–470, 2015.
You can also read