Keywords
patients - chatbot interaction - implementation - digital - social determinants of
health
Background and Significance
Background and Significance
There is growing interest in screening for social needs to understand and address
the link between health inequities and social determinants of health (SDoH),[1]
[2]
[3]
[4]
[5] the conditions in which people are born, grow, live, work, and age. Social needs
are the needs of an individual as a result of their SDoH, such as housing instability,
food insecurity, or unemployment.[6] Hospital emergency departments (EDs) may be one appropriate place for social needs
screening as EDs serve vulnerable populations with a high prevalence of social needs.[7]
[8] However, SDoH screening in the ED is not routine, and even when needs are identified,
referral to community services and follow-up may be beyond the current capacities
of many EDs.
Patients benefit from assistance to complete screening and contact community resources.[9] Yet implementing face-to-face social needs screening and referral in the ED is challenging
due to anticipated patient discomfort and clinician burden.[10]
[11] Self-administered screening could overcome these challenges. Potential approaches
include patient-facing surveys distributed via paper,[12]
[13]
[14] automated phone calls and text messaging,[15] tablets,[16]
[17]
[18] and personal health records.[19] While patient- and provider-facing SDoH screening tools exist, such as direct entry
by patients or providers into electronic health records (EHRs),[20] they face limited uptake. There are well-known disparities in patient adoption of
online portals and use of personal health records.[21]
[22] One reason for nonuse of patient portals includes privacy and information security
concerns, which indicates the importance of building patient trust in communication
systems.[23] Thus, getting patient input is important to address patients' health and social
needs[24]
[25]
[26]
[27]
[28] and to design more accessible and trustworthy approaches to better engage patients.
Chatbots are computer programs that simulate conversations with users that are increasingly
adopted in the health care field,[29]
[30] and have the potential to address the above screening barriers. Chatbots may increase
patient uptake as a more understandable and engaging tool compared to traditional
online surveys.[31]
[32]
[33] Further, patients may be more willing to disclose personal information[34] and social needs information[18] to a computer. Individuals may feel more inclined to use conversational agents for
discussing sensitive health topics, such as addiction,[35] depression,[36] and posttraumatic stress disorder,[37] as technology can enable more confidential methods of information and support seeking.[38]
[39] Despite growing interest in chatbot technology for health, there are few published
studies on conversational user interfaces in health care.[29]
[40]
[41]
[42] The literature on conversational agents in health care is largely aimed at treatment
and monitoring of health conditions, such as mental health,[43]
[44]
[45]
[46] Alzheimer's,[47] heart failure,[48] asthma,[49] and human immunodeficiency virus,[50] health service support, such as patient history taking,[51]
[52] and triage and diagnosis support,[53]
[54] and patient education, on topics such as sexual health,[55] smoking,[56] alcohol use,[57] and breast cancer.[58]
While studies of conversational agents in health care have shown moderate evidence
of usability and effectiveness,[42] there is a need for further exploration on the role of conversational agents in
real-world settings.[40] User feedback on conversational agents in health care remains mixed, with some users
expressing desire for interactivity and agent empathy, whereas others report a dislike
of these qualities.[32]
[33]
[42] There are also still few conversational agent evaluations with users in clinical
settings.[40]
[59] Prior research on conversational agents in real-world settings has identified the
need for providing actionable and accurate information.[59] Additionally, prior work that has studied conversational agents for social needs
screening in clinical settings has indicated the importance of designing for and with
vulnerable populations, such as people with low health literacy, to improve chatbot
understandability.[32] More work is needed to evaluate chatbots in clinical settings, in particular chatbots
for social needs screening. We have little knowledge about patients' sharing practices
around social needs-related data during real-world clinic visits. Furthermore, we
have not established if patients find chatbots feasible and acceptable for social
needs screening, nor whether the ED is an appropriate site for social needs screening
chatbots. This is important as it could lead to a screening process that is more likely
to result in patients receiving care for social needs. We build on prior work to evaluate
a chatbot implementation situated within the ED workflow, and investigate patient
perceptions of the screening and resource provision in the ED context.
Objectives
We investigated three implementation outcome measures[60] to evaluate the success of a chatbot implementation for social needs screening at
a large hospital ED. Our aim was to address the following research questions: (1)
How do patients rate the acceptability, feasibility, and appropriateness of a chatbot
implementation in the ED for social needs screening? (2) What are patient perceptions
of using a chatbot for social needs screening?
Methods
Study Design
In this study, we deployed a chatbot for social needs screening in a real-world context
to understand patients' perspectives on the acceptability, feasibility, and appropriateness
of using the tool in the ED. We used concurrent triangulation as a mixed-methods approach
to confirm and corroborate findings within our study.[61] First, we collected ratings of implementation measures via surveys with participants
who completed screening using the chatbot. Second, we conducted follow-up interviews
with a subset of participants to further understand patient perspectives.
Setting and Recruitment
The study took place in the ED at Harborview Medical Center, a large, public, tertiary
care teaching hospital, in the Pacific Northwest region of the United States from
November 9, 2020, to February 28, 2021. Patients were approached by a research assistant
after completing ED registration and triage. They were considered eligible if they
were at least 18 years old, English or Spanish-speaking, and did not have an acute
medical or psychiatric condition. We used the Emergency Severity Index (ESI) as the
qualification for identifying patients who would be able to participate in the study.[62] Patients were considered eligible if they had an ESI of 3 to 5 (i.e., not requiring
immediate medical attention based on triage algorithm). In the chatbot screening,
participants read a short introduction to the study and were asked if they consent
to participating by clicking “Okay, let's start” to proceed.
Collection of Social Needs and Implementation Measures
The chatbot for social needs screening provides relevant community resources to ED
patients ([Figs. 1] and [2]). Participants interacted with the chatbot on an iPad and could use optional disposable
headphones. The screening was available in English and Spanish. Participants used
the chatbot to answer 16 questions about their social needs that were adapted from
the Accountable Health Communities Health-Related Social Needs (AHC HRSN) Screening
Tool,[63] the Benefits Eligibility Screening Tool (BEST),[64] and the Los Angeles County Health Agency (LACHA) screening guide (Johnson 2019[65]; see [Supplementary Appendix: Screening Questionnaire], available in the online version).
Fig. 1 Screenshots of user interaction with HarborBot for social needs screening.
Fig. 2 Screenshots of chatbot screening output with user responses and list of tailored
community resources.
At the end of the screening, the chatbot asked participants to rate three implementation
outcome measures[60] to assess the acceptability, feasibility, and appropriateness of the chatbot on
a Likert scale from 1 “completely disagree” to 5 “completely agree.” Using these measures,
“acceptability” assesses the perception that a given innovation is agreeable or satisfactory,
“appropriateness” assesses the perceived compatibility of the innovation for a given
issue and practice setting, and “feasibility” assesses the extent to which the innovation
can be successfully used or carried out.[60]
Participants were also asked 6 demographic questions about their age, gender, race/ethnicity,
education, relationship status, and insurance status. Finally, participants were asked
if they would be willing to take part in a follow-up interview. Participants were
eligible for a follow-up interview if they had a working phone number. Upon completion
of the screening, the participant was handed a printed copy of their responses and
a list of matching community resources ([Fig. 2]), and encouraged to share their responses with their ED care team. Participants
could optionally send their responses and resource list to themselves via email and
text.
Chatbot Design
HarborBot is a web application that is accessible on mobile phones and desktops (see
[Supplementary Appendix: Chatbot Design], available in the online version). The chatbot interacts with users through chat
and voice (output only) in a scripted dialogue. The front-end web application is hosted
on Google Cloud, developed using HTML, CSS, and Javascript, and uses Python to communicate
with multiple API services. [Fig. 1] shows the graphical user interface for HarborBot. We used BotUI (https://botui.org/), a Javascript framework, to build the chatbot user interface, and REDCap database[66] to store user responses. After screening completion, social needs are highlighted
in red and relevant resources are brought to the top of the page, but all the resources
were included to ensure that participants had access, regardless of whether or not
they chose to disclose their social needs. [Fig. 2] shows the graphical user interface for HarborBot's response summary and resource
page. We compiled a list of local community resource organizations to share with users
based upon resources distributed by social workers at the Harborview Medical Center
ED. These resources were drawn from the Emerald City Resource Guide[67] and Washington 211,[68] online databases that help connect people to community resources in Seattle and
Washington state. We followed the BEST, LACHA, and AHC HRSN Screening Tool's scoring
instructions on what responses constitute a social need for each domain, which then
determined if the corresponding resource is highlighted on the page.
Follow-up Interview
We interviewed participants about their experience using the chatbot. Participants
were contacted via email or text message accompanied by a phone call 2 to 4 days after
their ED visit. The follow-up interview was either conducted at the time of contact
or scheduled for a later date. The interviews were conducted by phone and were audio-recorded,
except for one participant who did not consent to be recorded. These interviews were
semi-structured and asked participants about their perceptions of whether the chatbot
was an acceptable, feasible, and appropriate way of screening (see [Supplementary Appendix: Qualitative Interview], available in the online version). We also asked participants how they used the
resource list, how they currently search for and access community resources, and in
what ways a chatbot could facilitate this process. The Health Literacy Single-Item
Literacy Screener (SILS) is a single-item question that was administered to identify
adults with limited reading ability.[69] Participants were offered a USD 30 gift card after the interview.
Data Analysis
We used descriptive statistics to analyze the participant demographic information
([Table 1]) and implementation ratings ([Table 2]). Analyses were performed using Microsoft Excel (version 16.43) and RStudio (version
2022.12.0 + 353). We followed an inductive-deductive thematic approach[70] in the analysis of the interview data. Three team members performed inductive coding
on an initial set of three interviews. Four team members then clustered the codes
to develop a codebook. We incorporated concepts from the Consolidated Framework for
Implementation Research framework[71] to draw from established concepts in implementation theory. Once all four team members
reached agreement on the codes, we applied the codebook to the remaining interviews.
Table 1
Study participant demographics
|
Screened participants (n = 350), n %
|
Interview participants (n = 22), n %
|
Age (y)
|
18–25
|
29 (8.3)
|
2 (9.1)
|
26–35
|
83 (23.7)
|
6 (27.3)
|
36–45
|
57 (16.3)
|
4 (18.2)
|
46–55
|
36 (10.3)
|
3 (13.6)
|
56–65
|
23 (6.6)
|
2 (9.1)
|
≥ 66
|
20 (5.7)
|
1 (4.5)
|
Prefer not to answer
|
102 (29.1)
|
4 (18.2)
|
Gender
|
Male
|
187 (53.4)
|
11 (50.0)
|
Female
|
135 (38.6)
|
11 (50.0)
|
Additional gender category
|
16 (4.6)
|
0 (0.0)
|
Prefer not to answer
|
12 (3.4)
|
0 (0.0)
|
Racial/ethnic background
|
White
|
134 (38.3)
|
11 (50.0)
|
Black, African American or African
|
78 (22.3)
|
4 (18.2)
|
Latin American, Central American, Mexican or Mexican American, Hispanic or Chicano
|
53 (15.1)
|
2 (9.1)
|
More than one race
|
38 (10.9)
|
2 (9.1)
|
Asian: Indian, Chinese, Filipino, Japanese, Korean, Vietnamese, Other
|
15 (4.3)
|
2 (9.1)
|
Other
|
14 (4.0)
|
0 (0.0)
|
Prefer not to answer
|
18 (5.1)
|
1 (4.5)
|
Education
|
Some college
|
83 (23.7)
|
5 (22.7)
|
High school graduate
|
72 (20.6)
|
4 (18.2)
|
Bachelor's degree
|
38 (10.9)
|
6 (27.3)
|
Less than high school
|
36 (10.3)
|
2 (9.1)
|
Some high school
|
31 (8.9)
|
2 (9.1)
|
Graduate school
|
31 (8.9)
|
1 (4.5)
|
Associate degree
|
28 (8.0)
|
2 (9.1)
|
Prefer not to answer
|
31 (8.9)
|
0 (0.0)
|
Relationship status
|
Single/never married
|
158 (45.1)
|
11 (50.0)
|
Married
|
62 (17.7)
|
1 (4.5)
|
Divorced
|
47 (13.4)
|
7 (31.8)
|
Committed relationship/partnered
|
30 (8.6)
|
3 (13.6)
|
Separated
|
15 (4.3)
|
0 (0.0)
|
Widowed
|
8 (2.3)
|
0 (0.0)
|
Prefer not to answer
|
30 (8.6)
|
0 (0.0)
|
Health insurance
|
Medicaid
|
90 (25.7)
|
5 (22.7)
|
No health insurance
|
56 (16.0)
|
2 (9.1)
|
Employer provided
|
53 (15.1)
|
3 (13.6)
|
Medicare
|
53 (15.1)
|
5 (22.7)
|
Don't know
|
28 (8.0)
|
4 (18.2)
|
Other
|
24 (7.0)
|
2 (9.1)
|
Charity care
|
8 (2.3)
|
0 (0.0)
|
Private health insurance
|
8 (2.3)
|
0 (0.0)
|
COBRA
|
1 (0.3)
|
1 (4.6)
|
Prefer not to answer
|
29 (8.3)
|
0 (0.0)
|
Health Literacy Single-Item Literacy Screener (SILS)
|
1: never
|
–
|
10 (45.5)
|
2: rarely
|
–
|
9 (40.9)
|
3: sometimes
|
–
|
3 (13.6)
|
4: often
|
–
|
0 (0.0)
|
5: always
|
–
|
0 (0.0)
|
Table 2
Ratings of implementation measures
Constructs
|
Implementation outcome measures
|
Sample size of respondents, n (%)
|
Average rating on 1–5 Likert's scale (SD)
|
Median rating on 1–5 Likert's scale (IQR)
|
Acceptability
|
I like the use of this chatbot to answer these questions
|
297 (84.9)
|
3.93 (0.98)
|
4 (1)
|
Feasibility
|
Using this chatbot to answer these questions seems easy to use
|
301 (86.0)
|
4.20 (0.86)
|
4 (1)
|
Appropriateness
|
Using this chatbot to answer these questions seems suitable
|
302 (86.3)
|
4.10 (0.86)
|
4 (1)
|
Abbreviations: IQR, interquartile range: SD, standard deviation.
Transcript coding was divided among the four team members, and during each iteration
of coding, team members coded one to two different transcripts. In research meetings,
questions or concerns related to particular excerpts were discussed. Each team member
reviewed the transcripts, and disagreements were discussed to achieve consensus. We
returned to the initial interviews to recode them with the finalized codebook. We
continued discussions across all the interviews to identify themes and patterns in
the interviews to explain the ratings and provide additional insights.
Results
Participant Characteristics
A total of 832 patients were approached and 410 patients (49%) agreed to participate
in the study. Of those who agreed, 353 patients completed the screening and 3 patients
under the age of 18 were removed. There were 350 participants who consented and completed
the screening. The participants who completed screening (“screened participants”)
ranged in age from 18 to 90 years old (mean 40.7, standard deviation [SD] = 14.7)
and were diverse in age, race/ethnicity, education, and insurance status, and nearly
half were single or never married ([Table 1]). Among the participants, 329 participants completed the screening in English and
21 participants in Spanish. The screening took 10.92 minutes on average (SD = 7.50).
Of the 350 participants, 22 agreed to follow-up interviews. We conducted follow-up
phone interviews and qualitative analysis concurrently until reaching thematic saturation.[72] Interview participants (P1–P22) ranged in age from 18 to 68 years old (mean 40.6,
SD = 14.4). They were largely representative of the demographics in the screened participant
sample, with a larger representation of White/Caucasian participants and smaller representation
of those who received some college or less. Three interview participants (13.6%) reported
that they “sometimes” need help to read written health material. The interviews lasted
on average 42 minutes.
RQ1: Patient Ratings of the Chatbot Implementation: Acceptability, Feasibility, and
Appropriateness
Our findings demonstrate the value of the chatbot which was rated by participants
as an acceptable, feasible, and appropriate means of social needs screening, with
average ratings of 3.93 (SD = 0.99), 4.20 (SD = 0.86), and 4.10 (SD = 0.86), respectively
([Table 2]). [Fig. 3] shows the Likert scale rating distribution for acceptability, feasibility, and appropriateness
of the chatbot. The majority of participants agreed that they liked using the chatbot
and it was easy to use and appropriate, with some discrepancy among the acceptability
ratings ([Fig. 3]). [Figs. 4 ] to [ 6] and [Tables 3] to [5] show the Likert rating response distribution by age, ethnicity, and education. There
were some differences in perceptions of acceptability between age groups, ethnicities,
and education levels. The percentage of the participants who agreed or completely
agreed that the chatbot is acceptable was 88.9% among younger participants aged 18
to 25, compared to 65.0% among participants more than 66 years old. Additionally,
79.7% of Black, African American, or African participants agreed or completely agreed
that the chatbot is acceptable, compared to 53.9 and 61.5% of Asian participants and
Other participants (who identified as Native American, Pacific Islander, or Middle
Eastern). Participants who completed less than high school, some college, or were
a high school graduate, 78.1, 79.5 and 77.3% respectively, agreed or completely agreed
that the chatbot is acceptable to a greater extent than participants in graduate school
or who completed some high school, 66.7 and 66.7%.
Fig. 3 Diverging stacked bar chart of Likert scale ratings for acceptability, feasibility,
and appropriateness, accompanied by mean and standard deviation for each measure.
The percentage of positive responses (agree and completely agree) is stacked on the
right and the percentage of negative responses (disagree and completely disagree)
is stacked on the left, with neutral (neither agree nor disagree) in the center.
Fig. 4 Diverging stacked bar charts of Likert's scale ratings for acceptability, feasibility,
and appropriateness with response distributions by age. The mean and standard deviation
for each group are shown on the right.
Fig. 5 Diverging stacked bar charts of Likert's scale ratings for acceptability, feasibility,
and appropriateness with response distributions by ethnicity. The mean and standard
deviation for each group are shown on the right.
Fig. 6 Diverging stacked bar charts of Likert's scale ratings for acceptability, feasibility,
and appropriateness with response distributions by education. The mean and standard
deviation for each group are shown on the right.
Table 3
Response distributions of acceptability ratings by age, ethnicity, and education
Acceptability
|
Completely disagree (%)
|
Disagree (%)
|
Neither agree nor disagree (%)
|
Agree (%)
|
Completely agree (%)
|
Average rating on 1–5 Likert's scale (SD)
|
All participants
|
13 (4.4)
|
8 (2.7)
|
53 (17.9)
|
136 (45.8)
|
87 (29.3)
|
3.93 (0.98)
|
Age (y)
|
18–25
|
0 (0.0)
|
0 (0.0)
|
3 (11.1)
|
16 (59.3)
|
8 (29.6)
|
4.19 (0.61)
|
26–35
|
4 (5.2)
|
2 (2.6)
|
16 (20.8)
|
35 (45.5)
|
20 (26.0)
|
3.84 (1.01)
|
36–45
|
1 (1.9)
|
4 (7.7)
|
8 (15.4)
|
24 (46.2)
|
15 (28.9)
|
3.92 (0.96)
|
46–55
|
3 (9.7)
|
0 (0.0)
|
4 (12.9)
|
11 (35.5)
|
13 (41.9)
|
4.0 (1.19)
|
56–65
|
1 (5.0)
|
2 (10.0)
|
2 (10.0)
|
11 (55.0)
|
4 (20.0)
|
3.75 (1.04)
|
>= 66
|
1 (5.0)
|
0 (0.0)
|
6 (30.0)
|
7 (35.0)
|
6 (30.0)
|
3.85 (1.01)
|
Prefer not to answer
|
3 (4.3)
|
0 (0.0)
|
14 (20.0)
|
32 (45.7)
|
21 (30.0)
|
3.97 (0.94)
|
Racial/ethnic background
|
White
|
3 (2.5)
|
1 (0.8)
|
24 (20.2)
|
57 (47.9)
|
34 (28.6)
|
3.99 (0.86)
|
Black, African American or African
|
5 (7.8)
|
1 (1.6)
|
7 (10.9)
|
29 (45.3)
|
22 (34.4)
|
3.97 (1.10)
|
Latin American, Central American, Mexican or Mexican American, Hispanic or Chicano
|
3 (5.9)
|
2 (3.9)
|
7 (13.7)
|
25 (49.0)
|
14 (27.5)
|
3.88 (1.04)
|
More than one race
|
0 (0.0)
|
1 (3.3)
|
6 (20.0)
|
14 (46.7)
|
9 (30.0)
|
4.03 (0.80)
|
Asian: Indian, Chinese, Filipino, Japanese, Korean, Vietnamese, Other
|
1 (7.7)
|
2 (15.4)
|
3 (23.1)
|
2 (15.4)
|
5 (38.5)
|
3.62 (1.33)
|
Other
|
1 (7.7)
|
1 (7.7)
|
3 (23.1)
|
6 (46.2)
|
2 (15.4)
|
3.54 (1.08)
|
Prefer not to answer
|
0 (0.0)
|
0 (0.0)
|
3 (42.9)
|
3 (42.9)
|
1 (14.3)
|
3.71 (0.70)
|
Education
|
Some college
|
2 (2.7)
|
0 (0.0)
|
13 (17.8)
|
38 (52.1)
|
20 (27.4)
|
4.01 (0.84)
|
High school graduate
|
3 (4.6)
|
2 (3.0)
|
10 (15.2)
|
32 (48.5)
|
19 (28.8)
|
3.94 (0.98)
|
Bachelor's degree
|
1 (2.6)
|
0 (0.0)
|
11 (29.0)
|
17 (44.7)
|
9 (23.7)
|
3.87 (0.86)
|
Less than high school
|
3 (9.4)
|
0 (0.0)
|
4 (12.5)
|
11 (34.4)
|
14 (43.8)
|
4.03 (1.19)
|
Some high school
|
4 (14.8)
|
3 (11.1)
|
2 (7.4)
|
10 (37.0)
|
8 (29.6)
|
3.56 (1.40)
|
Graduate school
|
0 (0.0)
|
2 (7.4)
|
7 (25.9)
|
11 (40.7)
|
7 (25.9)
|
3.85 (0.89)
|
Associate degree
|
0 (0.0)
|
1 (3.6)
|
6 (21.4)
|
12 (42.9)
|
9 (32.1)
|
4.04 (0.82)
|
Prefer not to answer
|
0 (0.0)
|
0 (0.0)
|
0 (0.0)
|
5 (83.3)
|
1 (16.7)
|
4.17 (0.37)
|
Abbreviation: SD, standard deviation.
Table 4
Response distributions of feasibility ratings by age, ethnicity, and education
Feasibility
|
Completely disagree (%)
|
Disagree (%)
|
Neither agree nor disagree (%)
|
Agree (%)
|
Completely agree (%)
|
Average rating on 1–5 Likert's scale (SD)
|
All participants
|
10 (3.3)
|
4 (1.3)
|
16 (5.3)
|
158 (52.5)
|
113 (37.5)
|
4.20 (0.86)
|
Age (y)
|
18–25
|
0 (0.0)
|
0 (0.0)
|
1 (3.7)
|
14 (51.9)
|
12 (44.4)
|
4.41 (0.56)
|
26–35
|
1 (1.3)
|
0 (0.0)
|
3 (3.8)
|
45 (57.0)
|
30 (38.0)
|
4.30 (0.66)
|
36–45
|
2 (3.8)
|
1 (1.9)
|
2 (3.8)
|
26 (49.1)
|
22 (41.5)
|
4.23 (0.90)
|
46–55
|
0 (0.0)
|
0 (0.0)
|
2 (6.3)
|
15 (46.9)
|
15 (46.9)
|
4.41 (0.61)
|
56–65
|
1 (5.6)
|
1 (5.6)
|
1 (5.6)
|
10 (55.6)
|
5 (27.8)
|
3.94 (1.03)
|
≥ 66
|
1 (5.0)
|
1 (5.0)
|
1 (5.0)
|
10 (50.0)
|
7 (35.0)
|
4.05 (1.02)
|
Prefer not to answer
|
5 (6.9)
|
1 (1.4)
|
6 (8.3)
|
38 (52.8)
|
22 (30.6)
|
3.99 (1.03)
|
Racial/ethnic background
|
White
|
2 (1.7)
|
2 (1.7)
|
8 (6.7)
|
64 (53.3)
|
44 (36.7)
|
4.22 (0.78)
|
Black, African American or African
|
5 (7.7)
|
0 (0.0)
|
3 (4.6)
|
29 (44.6)
|
28 (43.1)
|
4.15 (1.07)
|
Latin American, Central American, Mexican or Mexican American, Hispanic or Chicano
|
2 (4.0)
|
1 (2.0)
|
1 (2.0)
|
28 (56.0)
|
18 (36.0)
|
4.18 (0.89)
|
More than one race
|
0 (0.0)
|
1 (3.2)
|
2 (6.5)
|
15 (48.4)
|
13 (41.9)
|
4.29 (0.73)
|
Asian: Indian, Chinese, Filipino, Japanese, Korean, Vietnamese, Other
|
0 (0.0)
|
0 (0.0)
|
0 (0.0)
|
8 (61.5)
|
5 (38.5)
|
4.38 (0.49)
|
Other
|
1 (7.7)
|
0 (0.0)
|
2 (15.4)
|
7 (53.9)
|
3 (23.1)
|
3.85 (1.03)
|
Prefer not to answer
|
0 (0.0)
|
0 (0.0)
|
0 (0.0)
|
7 (77.8)
|
2 (22.2)
|
4.22 (0.42)
|
Education
|
Some college
|
1 (1.3)
|
0 (0.0)
|
3 (4.0)
|
42 (56.0)
|
29 (38.7)
|
4.31 (0.67)
|
High school graduate
|
4 (6.0)
|
0 (0.0)
|
4 (6.0)
|
34 (50.8)
|
25 (37.3)
|
4.13 (0.98)
|
Bachelor's degree
|
0 (0.0)
|
2 (5.3)
|
2 (5.3)
|
19 (50.0)
|
15 (39.5)
|
4.24 (0.78)
|
Less than high school
|
0 (0.0)
|
0 (0.0)
|
3 (9.4)
|
13 (40.6)
|
16 (50.0)
|
4.41 (0.65)
|
Some high school
|
1 (3.6)
|
2 (7.1)
|
0 (0.0)
|
16 (57.1)
|
9 (32.1)
|
4.07 (0.96)
|
Graduate school
|
2 (7.1)
|
0 (0.0)
|
2 (7.1)
|
15 (53.6)
|
9 (32.1)
|
4.04 (1.02)
|
Associate degree
|
0 (0.0)
|
0 (0.0)
|
2 (7.4)
|
15 (55.6)
|
10 (37.0)
|
4.30 (0.60)
|
Prefer not to answer
|
2 (33.3)
|
0 (0.0)
|
0 (0.0)
|
4 (66.7)
|
0 (0.0)
|
3.0 (1.41)
|
Abbreviation: SD, standard deviation.
Table 5
Response distributions of appropriateness ratings by age, ethnicity, and education
Appropriateness
|
Completely disagree (%)
|
Disagree (%)
|
Neither agree nor disagree (%)
|
Agree (%)
|
Completely agree (%)
|
Average rating on 1–5 Likert's scale (SD)
|
All participants
|
6 (2.0)
|
9 (3.0)
|
35 (11.6)
|
150 (49.7)
|
102 (33.8)
|
4.10 (0.86)
|
Age
|
|
|
|
|
|
|
18–25
|
0 (0.0)
|
0 (0.0)
|
4 (14.8)
|
15 (55.6)
|
8 (29.6)
|
4.15 (0.65)
|
26–35
|
1 (1.3)
|
1 (1.3)
|
8 (10.3)
|
45 (57.7)
|
23 (29.5)
|
4.13 (0.74)
|
36–45
|
2 (3.6)
|
2 (3.6)
|
4 (7.3)
|
26 (47.3)
|
21 (38.2)
|
4.13 (0.95)
|
46–55
|
0 (0.0)
|
2 (6.5)
|
2 (6.5)
|
14 (45.2)
|
13 (41.9)
|
4.23 (0.83)
|
56–65
|
0 (0.0)
|
2 (10.5)
|
0 (0.0)
|
12 (63.2)
|
5 (26.3)
|
4.05 (0.83)
|
≥ 66
|
0 (0.0)
|
1 (5.0)
|
5 (25.0)
|
5 (25.0)
|
9 (45.0)
|
4.10 (0.94)
|
Prefer not to answer
|
3 (4.2)
|
1 (1.4)
|
12 (16.7)
|
33 (45.8)
|
23 (31.9)
|
4.0 (0.96)
|
Racial/ethnic background
|
White
|
1 (0.9)
|
2 (1.7)
|
15 (12.7)
|
56 (47.5)
|
44 (37.3)
|
4.19 (0.78)
|
Black, African American or African
|
3 (4.6)
|
4 (6.2)
|
4 (6.2)
|
31 (47.7)
|
23 (35.4)
|
4.03 (1.04)
|
Latin American, Central American, Mexican or Mexican American, Hispanic or Chicano
|
1 (1.9)
|
2 (3.9)
|
5 (9.6)
|
29 (55.8)
|
15 (28.9)
|
4.06 (0.84)
|
More than one race
|
0 (0.0)
|
1 (3.1)
|
3 (9.4)
|
15 (46.9)
|
13 (40.6)
|
4.25 (0.75)
|
Asian: Indian, Chinese, Filipino, Japanese, Korean, Vietnamese, Other
|
0 (0.0)
|
0 (0.0)
|
3 (23.1)
|
5 (38.5)
|
5 (38.5)
|
4.15 (0.77)
|
Other
|
1 (7.1)
|
0 (0.0)
|
4 (28.6)
|
8 (57.1)
|
1 (7.1)
|
3.57 (0.90)
|
Prefer not to answer
|
0 (0.0)
|
0 (0.0)
|
1 (12.5)
|
6 (75.0)
|
1 (12.5)
|
4.0 (0.50)
|
Education
|
Some college
|
2 (2.7)
|
2 (2.7)
|
5 (6.7)
|
38 (50.7)
|
28 (37.3)
|
4.17 (0.87)
|
High school graduate
|
2 (3.0)
|
1 (1.5)
|
5 (7.6)
|
39 (59.1)
|
19 (28.8)
|
4.09 (0.83)
|
Bachelor's degree
|
0 (0.0)
|
1 (2.6)
|
8 (21.1)
|
19 (50.0)
|
10 (26.3)
|
4.0 (0.76)
|
Less than high school
|
0 (0.0)
|
1 (3.0)
|
5 (15.2)
|
10 (30.3)
|
17 (51.5)
|
4.30 (0.83)
|
Some high school
|
1 (3.5)
|
2 (6.9)
|
1 (3.5)
|
17 (58.6)
|
8 (27.6)
|
4.0 (0.95)
|
Graduate school
|
1 (3.6)
|
0 (0.0)
|
6 (21.4)
|
11 (39.3)
|
10 (35.7)
|
4.04 (0.94)
|
Associate degree
|
0 (0.0)
|
1 (3.7)
|
4 (14.8)
|
12 (44.4)
|
10 (37.0)
|
4.15 (0.80)
|
Prefer not to answer
|
0 (0.0)
|
1 (16.7)
|
1 (16.7)
|
4 (66.7)
|
0 (0.0)
|
3.50 (0.76)
|
Abbreviation: SD, standard deviation.
RQ2: Patients' Perceptions of Using the Chatbot for Social Needs Screening
Analysis of the interviews identified six qualitative themes that describe ways in
which participants perceived the chatbot as acceptable, feasible, and appropriate,
and potential barriers to use.
Acceptability
Participants were satisfied that the chatbot provided a responsive interaction which acknowledged patients' answers and replied with personalized resources.
Additionally, they liked how the chatbot afforded privacy during information disclosure, but raised questions about the security of their data.
Participants appreciated the chatbot screening as an important first step in fostering
a sense of care at the ED, while noting that it is important to follow up with patients
to ensure they access resources.
Chatbot Provides Responsive, Engaging Interaction
Overall, participants found that the chatbot was responsive and engaged them during
screening. Participants liked that the chatbot maintained the responsiveness of a
human interaction and guided them through each question ([Table 6]). Participants also liked that the chatbot provided personalized recommendations
for community resources, avoiding information overload through extraneous recommendations.
They appreciated that the conversation was brief, rather than repetitive, unlike past
surveys that asked many similar questions about the same type of social need. P13
was looking for food assistance and found that the resources were tailored to their
social needs ([Table 6]).
Table 6
Example quotes from the qualitative analysis
Theme
|
Theme description
|
Example quotes
|
Acceptability
|
Chatbot provides responsive, engaging interaction
|
“If you put yes or no, depending on what you put, you have another answer on the chatbot.
If [the screening] was just a couple of questions on paper, then you wouldn't receive
that reply” (P17).
“It seems more personal because it literally narrows down and takes out what you said
yes to, what you said no to. Then it only gives you information on what you need help
with, instead of giving you a load of information on certain things that [are not
relevant to you]...say, if you're not an alcoholic, it's not giving you a number to
AA … If you need a food bank, it's giving you a number to food banks, it's giving
you a number to donation places” (P13).
|
Acceptability
|
Chatbot helped preserve privacy during information disclosure, but prompted questions
about data sharing and security
|
“I'd rather answer my questions and everything with the chatbot. That way [other patients
are] not hearing what's going on with me as far as money. In fact, a lot of the things
that I don't like is I have to repeat, like give them my address, my phone number,
in front of these strangers who you don't know...I don't want to give out that kind
of information out loud...since we were online and the chatbot actually submitted
the information that I sent out, that was actually probably one of the things that
I felt safest with at that time” (P6).
“There's a couple times where I was in the ED by myself. I've been medicated with
morphine or something, and I have been out there and waiting for a cab, and this person
would sit next to me to try to grab whatever was in my bag” (P6).
“I have to try to be careful what app I'm using or whatever…because there are predators
out there that will steal your identity” (P10).
|
Feasibility
|
Chatbot is easy to use and understand
|
“The instructions were pretty self-explanatory. You could understand the instructions,
like when it was dragging you to the next page and what to do and all that. So that
was pretty cool that they broke everything down for you as you went along… I didn't
have to ask [the research assistant] anything the whole time I did it” (P5).
|
Feasibility
|
Chatbot screening is quick to complete
|
“I think you have some free time…[compared to] a survey that comes through email,
I know I get them all the time and almost never filled them out. So the way in which
the survey is administered [via chatbot], I think it's a good way to get more responses”
(P16).
|
Appropriateness
|
Chatbot screening is appropriate in the ED context
|
“I think in the setting of the hospital that it was easier to use the chatbot than
it would be to find a quiet place to sit down where you could have a discussion with
a person” (P1).
“I don't feel comfortable talking about my financial situation with my doctor...They
can be helpful if they provide you the information there, or they redirect you…Usually
there's no conversation to bring it up…Well, they're just telling you what to do to
make it better, or they're going to prescribe you something. So sometimes that conversation
doesn't go along with the housing” (P17).
“I don't want to tell them that I'm homeless because I feel like I'm being treated
differently as opposed if I just tell them, oh, okay, well I live over here… The whole
issue, I think just came down to, they found out I was homeless. I was sleeping outside.
The doctor expressed that they didn't want to do the surgery because I didn't have
a sterile place to heal. I said, well, that's what you guys are here for. You have
respite beds that you provide for people that need a place to heal. And so the answer
that I got was, well, we can't reserve respite beds” (P5).
|
Screening is the first step in fostering a sense of care
|
Screening is the first step in fostering a sense of care
|
“I just feel like the more access and the more ways of making people get the resources
the better. I don't feel like there should just be one way of getting resources out
to people…considering a lot of the day services will give you booklets with resources.
But the problem with that was the resources wouldn't be updated, so a lot of it was
outdated. A lot of places you would call were closed down. They wasn't operating no
more. So the booklet was useless at the end of the day” (P5).
“It's good to have an actual human being there...telling me that they want to get
you help or they can get you assistance, and then they stay there and you answer the
questions that they're asking you. It feels like a more believable situation...it
would be nice to maybe have someone contact you the day after you get out or a couple
of days after you get out and go over what you filled out instead of just an automated
voice” (P19).
“It was easy to answer because it had preloaded answers…but, [you] can't elaborate
too much with a chatbot…It asked, 'Are you or somebody in your household experiencing
hardship?' Then, I said yes. Then, it asks how, and I said income, but I wasn't able
to type in more than income…I wanted to say it was his income, not mine. That's the
issue right now because we're sitting here waiting for unemployment” (P6).
|
Chatbot Helped Preserve Privacy during Information Disclosure, but Prompted Questions
about Data Sharing and Security
Participants who did not want to be overheard in the ED valued the chatbot. They liked
that they could input their responses instead of speaking out loud ([Table 6]). There was a sense that the ED was not a secure place to discuss personal information
and the chatbot afforded privacy from answering questions in an open space. P6 was
not only worried about being overheard, but worried about other ED visitors who might
take and view their responses if they were on paper ([Table 6]). Privacy during information disclosure was very important to participants to avoid
direct judgment or stolen information. Participants desired that their information
be stored securely in the EHR after the chatbot interaction, and assumed that their
information would not be shared with unauthorized individuals. However, some participants
were cautious of what information to share with the chatbot as they felt it may lead
to stolen information. P10 was hesitant about sharing personal information via the
chatbot and explained that they try to be careful no matter what application they
use ([Table 6]). Together, these examples illustrate that participants found privacy-preserving
aspects of the chatbot to be acceptable, including no requirement to speak responses
out loud, and assurances that responses would not be shared inappropriately. However,
data security was a concern that reduced acceptability.
Feasibility
Participants found that the chatbot was a feasible method of social needs screening
in the ED. They found the chatbot easy to use, understand, and quick to complete.
Chatbot is Easy to Use and Understand
Participants found the chatbot easy to use which facilitated the successful completion
of screening. In support of their high ratings of feasibility, participants said they
could easily understand and answer the questions ([Table 6]). P13 agreed the chatbot was easy to use and compared the experience to playing
a computer game. Further, P5 liked using the tablet and selecting multiple choice
options rather than typing because their hand was broken. P10 described themselves
as less familiar with technology, but still found the chatbot as easy to use: “I don't
dislike it, but I'm just used to doing regular straight paper, not a tablet. I'm not
there yet…I'm not knowledgeable like some other people” (P10).
Chatbot Screening is Quick to Complete
When asked about how easy the chatbot was to use, participants found the chatbot feasible
because it could be used quickly and easily. The screening did not take a lot of time
to complete: “It was faster...more convenient maybe than talking to the representative
directly” (P3). The chatbot was direct and easy to understand, whereas people may
not be as direct: “You just answer Yes or No, it's not that difficult” (P2). P16 thought
it was an efficient and effective way to get responses since they had free time in
the ED waiting room and they would not be motivated to complete a survey sent via
email ([Table 6]). Overall, participants reported that they did not mind filling out questions to
pass the time and the chatbot only took a short time to complete.
Appropriateness
Participants perceived the chatbot as an appropriate technology for the setting. Participants
were comfortable sharing their social needs with the chatbot to avoid attention from
other ED visitors and social judgment present in face-to-face screening.
P1 found that the ED was busy and the chatbot was compatible with this context ([Table 6]). Most participants did not feel comfortable calling attention to themselves in
the ED, and using the chatbot on the tablet seemed like a casual, normal activity
that everyone was participating in. Participants cited fear of social judgment as
a reason that they preferred using the chatbot: “You might open up to a chatbot and
not a person…[there is] a lot of shame involved in some issues” (P4). P4 was searching
for stabilized housing options and had spent the last 15 years learning about homelessness.
Interacting with a chatbot has the potential to minimize social judgment that would
occur if talking with a health care worker “because you don't have to deal with its
[the chatbot's] attitude” (P6).
Participants also had different levels of comfort with what information to share with
health care providers. They may be uncomfortable or embarrassed to talk with a health
care provider about social needs, especially a provider they do not know. While P17
discussed how health care providers can be helpful to provide information about social
needs and redirect them to resources, they were not comfortable with bringing up their
social needs to their provider ([Table 6]). P21 even hesitated to disclose information, such as their ability to pay for utilities,
via the chatbot as they felt it may change the care they receive from ED providers.
Others discussed receiving lower quality care at the ED based on their social needs
in the past and did not want that to reoccur ([Table 6]).
Screening is the First Step in Fostering a Sense of Care
The chatbot was perceived as a valuable first step in learning about social resources.
P11 was homeless on and off for over 20 years and explained that screening for social
needs was important because “a lot of people don't know where to look…[and] don't
have access to the internet, so I think the way it [chatbot] was brought to me [on
a tablet] in the hospital was an awesome thing.” Even for those who know where to
look, using the chatbot was seen as another way of accessing information, particularly
since the current resources they are aware of may not be meeting their needs ([Table 6]). All participants said they would use the chatbot in the future and most were open
to tools that helped them discover resources.
However, effective follow-up on patients' social needs is necessary for patients to
feel cared for in the ED context. Participants mentioned that the screening should
feel personal and serve a purpose beyond collecting information. P19 felt the chatbot
did not provide personal benefits: “It was just a way of filling out the survey...It
didn't benefit anything really.” P19 wanted to have a person in the loop to ensure
that they are going to receive help ([Table 6]). Further, patients may want to elaborate on specific answers to ensure that they
get help ([Table 6]). For example, one participant tried to hand off the printed output to their provider,
but kept being redirected to the next staff person until they were able to share their
printed screening results with a social worker.
Participants rarely brought the printed responses and resource list to start a conversation
with their provider. Some participants were recurring patients who felt that ED providers
are very busy and did not want to bother them by bringing up their social needs. Although
few participants expressed concerns about sharing social needs through a chatbot in
the ED, the above-mentioned concerns and preferences around sharing social needs might
hinder some patients' sharing and early engagement with providers. To increase appropriateness
of a chatbot for social needs screening in an ED context, patients require secure
and reliable pathways for following up on resources.
Discussion
Our findings indicate that the chatbot implementation at the ED was perceived by patients
as a feasible, acceptable, and appropriate form of outreach that could increase uptake.
The ED has an explicit mission statement to care for vulnerable populations, and participants
recognized the ED as a place where many individuals with social needs go for assistance
and could participate in the screening. Those who may be more in need of resources,
such as those who have not completed an advanced degree, may be more receptive to
the chatbot screening, for example, patients who completed less than high school may
find the chatbot more acceptable than patients who completed graduate school. The
qualitative responses supported the survey responses when triangulating on the data.
This is significant, as it suggests that chatbots could facilitate a screening process
that ultimately connects patients to care for social needs, supporting the mission
of EDs as part of the social safety net and improving health and well-being for members
of the most vulnerable patient populations. Providers could use social needs information
to better personalize treatment plans and direct patients to resources available in
the hospital and community.
However, not all participants were positive about chatbots and strategies to improve
uptake in this group will be important future work. Those who did not want to use
the chatbot described themselves as being less familiar with new technology and applications.
The presence of a trained professional in the hospital ED can help to support the
screening process, in particular for older patients who may find a chatbot screening
less acceptable than younger patients. Some participants felt uncomfortable sharing
social needs with providers in the ED after completing the screening. This was due
to patients' perceived prioritization of medical needs over social needs at the ED,
and the potential negative impact on their emergency care. Although prior work indicated
that some patients want help with social needs from providers,[15] most interview participants did not discuss their screening results with ED providers.
For those who have data security concerns or do not want to discuss social needs with
their providers, future chatbot design should inform patients how their data will
be accessed for clinical purposes. If desired, they should be allowed to opt out of
data sharing. For patients who want to elaborate on their answers, they should be
provided flexibility within the chatbot interaction to express themselves and emphasize
what resource they need the most assistance with.
Some participants wanted reliable and actionable support in accessing resources, thus
one future direction is to link chatbots with existing health care systems to facilitate
referrals. It is important to establish pathways to alert providers to acute social
needs, get patients in touch with community-based organizations for resource referral,
and help providers follow up on patients afterwards. The design of a chatbot for social
needs screening may benefit from standardization since conversational user interfaces
in health care can lead to unintended consequences, such as miscommunication due to
information overload.[73] In the next steps, we plan to craft recommendations for system-wide implementation
of the screening and referral process developed. Further, departmental and health
system stakeholders plan to integrate social needs screening with existing technologies,
such as EHRs. There is ongoing research to prepopulate social needs by extracting
social needs-related information from clinical notes to address challenges of patient
data collection.[74]
The chatbot intervention could be further improved to reduce low uptake by establishing
trust through screening in additional contexts outside the ED. In future work, the
intervention could be evaluated at primary care clinics where some individuals may
feel more comfortable disclosing needs, such as community health clinics that serve
low-income patients. We believe that the ED waiting area is an ideal location for
social needs screening because idle time is spent there, many patients with social
needs are present, and the patients with lower ESI who would be more receptive to
participating make up the waiting room population. However, universal screening in
primary care may also be conducive to social needs screening, by supporting patient
comfort and promoting more regular social needs screenings. Prior research has indicated
there is little provider and patient discomfort with SDoH screening in primary care
settings[75]
[76] and that open discussions of social needs improved patients' relationships with
their health care team.[77] On the other hand, in EDs, providers have reported discomfort asking SDoH screening
questions they believed to be stigmatizing, and patients questioned the purpose of
the screening questions.[78] Although self-administered screening for social needs in primary care settings is
generally associated with high levels of acceptability by patients,[18]
[79] health care stakeholders have expressed concern about the presence of few patients
with social needs in primary care clinics which serve insured members who may be of
higher socioeconomic status.[80] Before implementing social needs screening interventions, primary care clinics should
evaluate their patient population to determine how they can reach patients facing
social needs.
Health interventions that have been proven to improve health outcomes are typically
longitudinal, tailored interventions that connect patients with community health workers
for case management.[81] While more institutional support is needed to follow-up with patients, chatbots
may serve as comfortable first touchpoint in the patient's journey through the ED
to disclose social needs. In many of the interview conversations, participants mentioned
their reasons for visiting the ED, including nonmedical issues, such as medical bill
assistance and medication refill. Additionally, some participants left the ED waiting
room before being admitted, due to long wait times. Thus, screening in the ED waiting
room prior to admission may have a wider reach and be completed by more individuals
than are actually admitted.
There are several limitations in this study. First, our findings are largely based
on participants' screening responses and interviews with a convenience sample. While
we aimed to recruit participants representative of the ED patient population, self-selection
bias may be present in participants who opted to participate in the study. For instance,
participants who had particularly negative experiences in the ED may be less prone
to participate or adopt chatbots. Second, the presence of the research team during
recruitment and novelty effect of the chatbot could also have influenced their use
and feedback on the chatbot. Finally, our study was conducted in a large public hospital
in one geographic region of the United States, which may limit the generalizability
of our findings. Despite these limitations, our study has a number of strengths, including
its reach and mixed-methods approach that provide important groundwork to guide future
studies.
Conclusion
We evaluated patients' perceptions of feasibility, acceptability, and appropriateness
of using a chatbot for social needs screening in the ED by collecting ratings and
conducting follow-up interviews among a diverse sample. Our findings demonstrate that
chatbots are an acceptable, feasible, and appropriate form of screening for patients
and can successfully engage a large, diverse patient population in the ED setting.
Participants observed that the chatbot screening was responsive, easy to use, efficient,
comfortable, and enhanced privacy during information disclosure. In future work, health
system stakeholders plan to integrate social needs screening with existing technologies,
such as EHRs, to augment patient data collection with clinical notes information,
and to reduce provider burden and information overload.
Clinical Relevance Statement
Clinical Relevance Statement
The chatbot screening has the potential to reduce ED provider and social worker burden
through EHR integration to summarize patients' acute social needs and automatic referral
to the relevant department. Providers may not discuss social needs with patients because
there is not an established pathway to address them. The chatbot screening can therefore
help to identify and address social needs that may go unaddressed during patient visits.
Without knowledge of patients' social needs, such as their inability to afford prescribed
medication, the effectiveness of health care can be diminished. Given that patients
may be concerned about social needs disclosure, health systems should facilitate social
needs screening to protect patient privacy and improve treatment.
Multiple-Choice Questions
Multiple-Choice Questions
-
When implementing a chatbot for social needs screening, which of the following are
important intervention qualities for users?
-
High noise level
-
Ease of use
-
Entertainment
-
Complex language
Correct Answer: The correct answer is option b because participants found the chatbot easy to use
which facilitated the successful completion of screening.
-
When implementing an intervention at the ED, which of the following help to facilitate
the intervention?
-
Number of patients
-
Patient hobbies
-
Time of day
-
Health care workers
Correct Answer: The correct answer is option d because health care workers, such as research assistants,
physicians, nurses, and social workers play a role in facilitating the screening intervention
in the ED waiting room and responding to patients who bring up their screening responses
and results.