S. L. Gable, H. T. Reis, E. A. Impett, and E. R. Asher, “What
do you do when things go right? The intrapersonal and
interpersonal benefits of sharing positive events,” Journal
of Personality and Social Psychology, vol. 87, no. 2,
pp. 228–245, 2004. DOI: 10.1037/0022-3514.87.2.228.
[2] J. Frattaroli, “Experimental disclosure and its moderators:
a meta-analysis,” Psychological Bulletin, vol. 132, no. 6,
pp. 823–865, 2006. DOI: 10.1037/0033-2909.132.6.823.
[3] P. G. Frisina, J. C. Borod, and S. J. Lepore, “A meta-analysis of
the effects of written emotional disclosure on the health
outcomes of clinical populations,” The Journal of Nervous
and Mental Disease, vol. 192, no. 9, pp. 629–634, 2004.
[Online]. DOI: 10.1097/01.nmd.0000138317.30764.63.
Available: https://journals.lww.com/jonmd/Fulltext/2004/
09000/A_Meta_Analysis_of_the_Effects_of_Written.8.aspx.
[4] E. Kennedy-Moore and J. C. Watson, “How and when does
emotional expression help?” Review of General Psychology,
vol. 5, no. 3, pp. 187–212, 2001. [Online].
DOI: 10.1037/1089-2680.5.3.187.
Available: https://doi.org/10.1037/1089-2680.5.3.187.
[5] D. M. Sloan, “Self-disclosure and psychological well-being,”
in Social psychological foundations of clinical psychology,
J. E. Maddux and J. P. Tangney, (Eds.), New York, NY, US:
The Guilford Press, 2010, pp. 212–225.
[6] V. J. Derlega, B. A. Winstead, R. J. Lewis, and J. Maddux,
“Clients’ responses to dissatisfaction in psychotherapy: a
test of Rusbult’s exit-voice-loyalty-neglect model,” Journal of
Social and Clinical Psychology, vol. 12, no. 3, pp. 307–318,
1993. DOI: 10.1521/jscp.1993.12.3.307.
[7] C. Breazeal, “Toward sociable robots,” Robotics and
Autonomous Systems, vol. 42, no. 3, pp. 167–175, 2003.
DOI: 10.1016/S0921-8890(02)00373-1.
[8] N. L. Robinson, T. V. Cottier, and D. J. Kavanagh,
“Psychosocial health interventions by social robots: systematic review of randomized controlled trials,” Journal of
Medical Internet Research, vol. 21, no. 5, e13203, 2019.
DOI: 10.2196/13203.
[9] A. A. J. Scoglio, E. D. Reilly, J. A. Gorman, and C. E. Drebing,
“Use of social robots in mental health and well-being research: systematic review,” Journal of Medical Internet
Research, vol. 21, no. 7, e13322, 2019. DOI: 10.2196/13322.
[10] E. Broadbent, K. Peri, N. Kerse, C. Jayawardena, I. H. Kuo,
C. Datta, and B. MacDonald, “Robots in older people’s homes
to improve medication adherence and quality of life: a randomised cross-over trial,” in Social Robotics, ICSR 2014,
Lecture Notes in Computer Science, vol. 8755, M. Beetz,
B. Johnston, and M.-A. Williams (Eds.), Cham: Springer
International Publishing, 2014, pp. 64–73,
DOI: 10.1007/978-3-319-11973-1_7.
[11] E. Broadbent, R. Tamagawa, A. Patience, B. Knock, N. Kerse,
K. Day, and B. A. MacDonald, “Attitudes towards health-care
robots in a retirement village,” Australasian Journal on
Ageing, vol. 31, no. 2, pp. 115–120, 2012.
DOI: 10.1111/j.1741-6612.2011.00551.x.
[12] N. Akalin, A. Kristoffersson, and A. Loutfi, “The influence of
feedback type in robot-assisted training,” Multimodal
154 Guy Laban et al.
Technologies and Interaction, vol. 3, no. 4, art. 67, 2019.
DOI: 10.3390/mti3040067.
[13] N. Akalin, A. Kiselev, A. Kristoffersson, and A. Loutfi,
“The relevance of social cues in assistive training with a
social robot,” in Social Robotics, ICSR 2018, Lecture
Notes in Computer Science, vol. 11357, S. S. Ge et al. (Eds.),
Cham: Springer International Publishing, 2018,
pp. 462–471.
[14] R. Feingold-Polak, A. Elishay, Y. Shahar, M. Stein, Y. Edan,
and S. Levy-Tzedek, “Differences between young and old
users when interacting with a humanoid robot: a qualitative
usability study,” Paladyn, Journal of Behavioral Robotics, vol.
9, no. 1, pp. 183–192, 2018. DOI: 10.1515/pjbr-2018-0013.
[15] R. F. Polak and S. L. Tzedek, “Social robot for rehabilitation:
expert clinicians and post-stroke Patients™ evaluation following a long-term intervention,” in Proceedings of the 2020
ACM/IEEE International Conference on Human-Robot
Interaction, HRI ‘20, New York, NY, USA: Association for
Computing Machinery, 2020, pp. 151–160.
DOI: 10.1145/3319502.3374797.
[16] E. Broadbent, “Interactions with robots: the truths we reveal
about ourselves,” Annual Review of Psychology, vol. 68,
no. 1, pp. 627–652, 2017.
DOI: 10.1146/annurevpsych-010416-043958.
[17] S. Jeong, D. E. Logan, M. S. Goodwin, S. Graca, B. O’Connell,
H. Goodenough, et al. “A social robot to mitigate stress,
anxiety, and pain in hospital pediatric care,” in Proceedings
of the Tenth Annual ACM/IEEE International Conference on
Human-Robot Interaction Extended Abstracts, HRI ‘15
Extended Abstracts, New York, NY, USA: Association for
Computing Machinery, 2015, pp. 103–104,
DOI: 10.1145/2701973.2702028. [Online]. Available:
https://doi.org/10.1145/2701973.2702028.
[18] A. Ostrowski, D. DiPaola, E. Partridge, H. W. Park, and
C. Breazeal, “Older adults living with social robots: promoting social connectedness in long-term communities,”
IEEE Robotics and Automation Magazine, vol. 26, no. 2,
pp. 59–70, 2019. DOI: 10.1109/MRA.2019.2905234.
[19] H. Robinson, B. MacDonald, N. Kerse, and E. Broadbent,
“The psychosocial effects of a companion robot: a randomized controlled trial,” Journal of the American Medical
Directors Association, vol. 14, no. 9, pp. 661–667, 2013.
DOI: 10.1016/j.jamda.2013.02.007.
[20] H. Ling and E. Björling, “Sharing stress with a robot: what
would a robot say?” Human-Machine Communication, vol. 1,
pp. 133–158, 2020. DOI: 10.30658/hmc.1.8.
[21] R. Yu, E. Hui, J. Lee, D. Poon, A. Ng, K. Sit, et al., “Use of a
therapeutic, socially assistive pet robot (PARO) in improving
mood and stimulating social interaction and communication
for people with dementia: study protocol for a randomized
controlled trial,” JMIR Research Protocols, vol. 4, no. 2, e45,
2015. DOI: 10.2196/resprot.4189.
[22] M. Petrovic and A. Gaggioli, “Digital mental health tools for
caregivers of older adults — a scoping review,” Frontiers in
Public Health, vol. 8, p. 128, 2020. [Online].
DOI: 10.3389/fpubh.2020.00128. Available: https://www.
frontiersin.org/article/10.3389/fpubh.2020.00128.
[23] H. L. Colquhoun, J. E. Squires, N. Kolehmainen, C. Fraser, and
J. M. Grimshaw, “Methods for designing interventions to
change healthcare professionals’ behaviour: a systematic
review,” Implementation Science, vol. 12, art. 30, 2017.
DOI: 10.1186/s13012-017-0560-5.
[24] D. Wight, E. Wimbush, R. Jepson, and L. Doi, “Six steps in
quality intervention development (6SQuID),” Journal of
Epidemiology and Community Health, vol. 70, pp. 520–525,
2016. DOI: 10.1136/jech-2015-205952.
[25] G. Riva, R. M. Baños, C. Botella, B. K. Wiederhold, and A.
Gaggioli, “Positive technology: using interactive technologies to promote positive functioning,” Cyberpsychology,
Behavior, and Social Networking, vol. 15, no. 2, pp. 69–77,
2012. DOI: 10.1089/cyber.2011.0139.
[26] E. S. Cross, R. Hortensius, and A. Wykowska, “From social
brains to social robots: applying neurocognitive insights to
human-robot interaction,” Philosophical Transactions of
the Royal Society B: Biological Sciences, vol. 374, no. 1771,
2019. DOI: 10.1098/rstb.2018.0024.
[27] A. Beck, Z. Yumak, and N. Magnenat-Thalmann, “Body
movements generation for virtual characters and social
robots,” in Social Signal Processing, J. K. Burgoon,
N. Magnenat-Thalmann, M. Pantic, and A. Vinciarelli, (Eds.),
Cambridge: Cambridge University Press, 2017, pp. 273–286.
DOI: 10.1017/9781316676202.020.
[28] R. Hortensius, F. Hekele, and E. S. Cross, “The perception of
emotion in artificial agents,” IEEE Transactions on Cognitive
and Developmental Systems, vol. 10, no. 4, pp. 852–864,
2018. DOI: 10.1109/TCDS.2018.2826921.
[29] G. Sandini, V. Mohan, A. Sciutti, and P. Morasso, “Social
cognition for human-robot symbiosis – challenges and
building blocks,” Frontiers in Neurorobotics, vol. 12, art. 34,
2018. DOI: 10.3389/fnbot.2018.00034.
[30] R. Hortensius and E. S. Cross, “From automata to animate
beings: the scope and limits of attributing socialness to artificial agents,” Annals of the New York Academy of Sciences,
vol. 1426, no. 1, pp. 93–110, 2018. DOI: 10.1111/nyas.13727.
[31] E. S. Cross, R. Ramsey, R. Liepelt, W. Prinz, and A. F. de C.
Hamilton, “The shaping of social perception by stimulus and
knowledge cues to human animacy,”Philosophical Transactions
of the Royal Society of London, Series B, Biological Sciences,
vol. 371, no. 1686, 2016. DOI: 10.1098/rstb.2015.0075.
[32] A. Klapper, R. Ramsey, D. Wigboldus, and E. S. Cross, “The
control of automatic imitation based on bottom–up and
top–down cues to animacy: insights from brain and behavior,” Journal of Cognitive Neuroscience, vol. 26, no. 11,
pp. 2503–2513, 2014. DOI: 10.1162/jocn_a_00651.
[33] G. Laban and T. Araujo, “Working together with conversational agents: the relationship of perceived cooperation with
service performance evaluations,” in Chatbot Research and
Design, CONVERSATIONS 2019, Lecture Notes in Computer
Science, vol. 11970, F. Asbjørn, et al., (Eds.), Cham: Springer
International Publishing, 2020, pp. 215–228.
DOI: 10.1007/978-3-030-39540-7_15.
[34] C. Özdem, E. Wiese, A. Wykowska, H. Müller, M. Brass, and
F. V. Overwalle, “Believing androids – fMRI activation in the
right temporo-parietal junction is modulated by ascribing
intentions to non-human agents,” Social Neuroscience,
vol. 12, no. 5, pp. 582–593, 2017. [Online].
DOI: 10.1080/17470919.2016.1207702. Available: https://
doi.org/10.1080/17470919.2016.1207702.
[35] R. L. Daft and R. H. Lengel, “Organizational information
requirements, media richness and structural design,”
Tell me more! Assessing interactions with social robots from speech 155
Management Science, vol. 32, no. 5, pp. 554–571, 1986.
DOI: 10.1287/mnsc.32.5.554.
[36] J. R. Carlson and R. W. Zmud, “Channel expansion theory and
the experiential nature of media richness perceptions,” The
Academy of Management Journal, vol. 42, no. 2, pp. 153–170,
1999. DOI: 10.2307/257090.
[37] H. Kreiner and Y. Levi-Belz, “Self-disclosure here and now:
combining retrospective perceived assessment with dynamic
behavioral measures,” Frontiers in Psychology, vol. 10,
art. 558, 2019. DOI: 10.3389/fpsyg.2019.00558. [Online].
Available: https://www.frontiersin.org/article/10.3389/
fpsyg.2019.00558.
[38] S. M. Jourard, Self-Disclosure: An Experimental Analysis of
the Transparent Self, Oxford, England: John Wiley, 1971.
[39] S. M. Jourard and P. Lasakow, “Some factors in selfdisclosure,” The Journal of Abnormal and Social Psychology,
vol. 56, no. 1, pp. 91–98, 1958. DOI: 10.1037/h0043357.
[40] J. H. Kahn, B. E. Hucke, A. M. Bradley, A. J. Glinski, and
B. L. Malak, “The distress disclosure index: a research review
and multitrait–multimethod examination,” Journal of
Counseling Psychology, vol. 59, no. 1, pp. 134–149, 2012.
DOI: 10.1037/a0025716.
[41] C. Antaki, R. Barnes, and I. Leudar, “Diagnostic formulations
in psychotherapy,” Discourse Studies, vol. 7, no. 6,
pp. 627–647, 2005. DOI: 10.1177/1461445605055420.
[42] J. Omarzu, “A disclosure decision model: determining how
and when individuals will self-disclose,” Personality and
Social Psychology Review, vol. 4, no. 2, pp. 174–185, 2000.
DOI: 10.1207/S15327957PSPR0402_05.
[43] Y. R. Tausczik and J. W. Pennebaker, “The psychological
meaning of words: LIWC and computerized text analysis
methods,” Journal of Language and Social Psychology, vol.
29, no. 1, pp. 24–54, 2010. DOI: 10.1177/0261927X09351676.
[44] B. Weiss, Talker Quality in Human and Machine Interaction:
Modeling the Listener’s Perspective in Passive and
Interactive Scenarios, Switzerland: Springer Nature
Switzerland AG, 2019. DOI: 10.1007/978-3-030-22769-2.
[45] R. W. Frick, “Communicating emotion: the role of prosodic
features,” Psychological Bulletin, vol. 97, no. 3, pp. 412–429,
1985. DOI: 10.1037/0033-2909.97.3.412.
[46] P. Roach, R. Stibbard, J. Osborne, S. Arnfield, and J. Setter,
“Transcription of prosodic and paralinguistic features of
emotional speech,” Journal of the International Phonetic
Association, vol. 28, no. 1–2, pp. 83–94, 1998.
DOI: 10.1017/S0025100300006277.
[47] Y. Yang, C. Fairbairn, and J. F. Cohn, “Detecting depression
severity from vocal prosody,” IEEE transactions on affective
computing, vol. 4, no. 2, pp. 142–150, Apr. 2013.
DOI: 10.1109/T-AFFC.2012.38.
[48] K. S. R., T. Johnstone, and G. Klasmeyer, “Vocal expression of
emotion,” in Series in affective science. Handbook of
affective sciences, R. J. Davidson, K. R. Scherer, and
H. H. Goldsmith (Eds.), New York, NY: Oxford University
Press, 2003, pp. 433–456.
[49] C. L. Giddens, K. W. Barron, J. Byrd-Craven, K. F. Clark, and
A. S. Winter, “Vocal indices of stress: a review,” Journal of
Voice, vol. 27, no. 3, pp. 21–390, May 2013.
DOI: 10.1016/j.jvoice.2012.12.010.
[50] R. Ruiz, C. Legros, and A. Guell, “Voice analysis to predict
the psychological or physical state of a speaker,” Aviation,
Space, and Environmental Medicine, vol. 61, no. 3,
pp. 266–271, 1990.
[51] G. M. Slavich, S. Taylor, and R. W. Picard, “Stress measurement using speech: recent advancements, validation issues,
and ethical and privacy considerations,” Stress, vol. 22, no. 4,
pp. 408–413, 2019. DOI: 10.1080/10253890.2019.1584180.
[52] M. van Puyvelde, X. Neyt, F. McGlone, and N. Pattyn, “Voice
stress analysis: a new framework for voice and effort in
human performance,” Frontiers in Psychology, vol. 9,
2018. DOI: 10.3389/fpsyg.2018.01994.
[53] Y. Levi-Belz and H. Kreiner, “What you say and how you say
it: analysis of speech content and speech fluency as predictors of judged self-disclosure,” Social Psychological and
Personality Science, vol. 7, no. 3, pp. 232–239, 2016.
DOI: 10.1177/1948550616632575. [Online]. Available:
10.1177/1948550616632575.
[54] A. M. Aroyo, F. Rea, G. Sandini, and A. Sciutti, “Trust and
social engineering in human robot interaction: will a robot
make you disclose sensitive information, conform to its recommendations or gamble?” IEEE Robotics and Automation
Letters, vol. 3, no. 4, pp. 3701–3708, 2018.
DOI: 10.1109/LRA.2018. 2856272.
[55] C. L. Bethel, M. R. Stevenson, and B Scassellati, “Secretsharing: Interactions between a child, robot, and adult,” in
2011 IEEE International Conference on Systems, Man, and
Cybernetics, Anchorage, AK, USA: IEEE, 2011,
pp. 2489–2494. DOI: 10.1109/ICSMC.2011.6084051.
[56] J.-J. De Groot, E. Barakova, T. Lourens, E. van Wingerden, and
P. Sterkenburg, “Game-based human-robot interaction promotes self-disclosure in people with visual impairments and
intellectual disabilities BT – understanding the brain function
and emotions,” in: J. M. Ferrández Vicente, J. R. ÁlvarezSánchez, F. de la Paz López, J. Toledo Moreo, and H. Adeli
(Eds.), Understanding the Brain Function and Emotions.
IWINAC 2019. Lecture Notes in Computer Science, vol. 11486,
Cham: Springer International Publishing, 2019, pp. 262–272.
[57] G. E. Birnbaum, M. Mizrahi, G. Hoffman, H. T. Reis, E. J. Finkel,
and O. Sass, “What robots can teach us about intimacy:
the reassuring effects of robot responsiveness to human
disclosure,” Computers in Human Behavior, vol. 63,
pp. 416–423, 2016. DOI: 10.1016/j.chb.2016.05.064.
[58] G. E. Birnbaum, M. Mizrahi, G. Hoffman, H. T. Reis, E. J. Finkel,
and O. Sass, “Machines as a source of consolation: robot
responsiveness increases human approach behavior and
desire for companionship,” in 2016 11th ACM/IEEE
International Conference on Human-Robot Interaction (HRI),
2016, pp. 165–172. DOI: 10.1109/HRI.2016.7451748.
[59] G. Hoffman, G. E. Birnbaum, K. Vanunu, O. Sass, and
H. T. Reis, “Robot responsiveness to human disclosure
affects social impression and appeal,” in Proceedings of
the 2014 ACM/IEEE International Conference on HumanRobot Interaction, HRI ‘14, New York, NY, USA: Association for
Computing Machinery, 2014, pp. 1–8.
DOI: 10.1145/2559636.2559660.
[60] H. Kumazaki, Z. Warren, A. Swanson, Y. Yoshikawa,
Y. Matsumoto, H. Takahashi, et al., “Can robotic systems
promote self-disclosure in adolescents with autism spectrum
disorder?” A Pilot Study, vol. 9, art. 36, 2018. [Online].
Available: https://www.frontiersin.org/article/10.3389/
fpsyt.2018.00036.
156 Guy Laban et al.
[61] E. A. Björling, E. Rose, A. Davidson, R. Ren, and D. Wong, “Can
we keep him forever? Teen’s engagement and desire for emotional connection with a social robot,” International Journal of
Social Robotics, vol. 12, pp. 65–77, 2019. [Online].
DOI: 10.1007/s12369-019-00539-6. Available: https://
doi.org/10.1007/s12369-019-00539-6.
[62] M. L. Traeger, S. S. Sebo, M. Jung, B. Scassellati, and
N. A. Christakis, “Vulnerable robots positively shape human
conversational dynamics in a human-robot team,” Proc. of
the National Academy of Sciences, vol. 117, no. 12,
pp. 6370–6375, 2020. DOI: 10.1073/pnas.1910402117.
[63] A. Ho, J. Hancock, and A. S. Miner, “Psychological, relational,
and emotional effects of self-disclosure after conversations
with a chatbot,” The Journal of Communication, vol. 68, no. 4,
pp. 712–733, 2018. DOI: 10.1093/joc/jqy026.
[64] D. L. Johanson, H. S. Ahn, B. A. MacDonald, B. K. Ahn, J. Lim,
E. Hwang, et al., “The effect of robot attentional behaviors on
user perceptions and behaviors in a simulated health care
interaction: randomized controlled trial,” Journal of Medical
Internet Research, vol. 21, no. 10, e13667, 2019.
DOI: 10.2196/13667.
[65] D. L. Johanson, S. A. Ho, C. J. Sutherland, B. Brown, B. A.
MacDonald, Y. L. Jong, et al., “Smiling and use of first-name
by a healthcare receptionist robot: effects on user perceptions, attitudes, and behaviours,” Paladyn, Journal of
Behavioral Robotics, vol. 11, no. 1, pp. 40–51, 2020.
DOI: 10.1515/pjbr-2020-0008.
[66] H. Kumazaki, Y. Yoshikawa, Y. Yoshimura, T. Ikeda, C.
Hasegawa, D. N. Saito, et al., “The impact of robotic intervention on joint attention in children with autism spectrum
disorders,” Molecular Autism, vol. 9, no. 46, 2018.
DOI: 10. 1186/s13229-018-0230-8.
[67] G. Laban and T. Araujo, “The effect of personalization techniques in users perceptions of conversational recommender
system,” in IVA ‘20: Proceedings of the 20th ACM
International Conference on Intelligent Virtual Agents
(IVA ‘20), Virtual Event, Glasgow, Scotland UK: ACM, 2020,
art. 34, pp. 1–3. DOI: 10.1145/3383652.3423890.
[68] Y.-C. Lee, N. Yamashita, Y. Huang, and W. Fu, ““I hear you,
I feel you”: Encouraging deep self-disclosure through a
chatbot,” in Proceedings of the 2020 CHI Conference on
Human Factors in Computing Systems, CHI ‘20, New York, NY,
USA: Association for Computing Machinery, 2020, pp. 1–12.
DOI: 10.1145/3313831.3376175.
[69] M. Shiomi, A. Nakata, M. Kanbara, and N. Hagita, “Robot
reciprocation of hugs increases both interacting times and
self-disclosures,” International Journal of Social Robotics,
2020. DOI: 10.1007/s12369-020-00644-x.
[70] G. Laban, V. Morrison, and E. S. Cross, “Let’s talk about it!
Subjective and objective disclosures to social robots,” in
Companion of the 2020 ACM/IEEE International Conference
on Human-Robot Interaction, Cambridge, UK: Association
for Computing Machinery, 2020, pp. 328–330.
DOI: 10.1145/3371382.3378252.
[71] L. D. Nelson, J. P. Simmons, and U. Simonsohn, “Let’s publish fewer papers,” Psychological Inquiry, vol. 23, no. 3,
pp. 291–293, 2012. DOI: 10.1080/1047840X.2012.705245.
[72] J. P. Simmons, L. D. Nelson, and U. Simonsohn, “False-positive psychology: undisclosed flexibility in data collection
and analysis allows presenting anything as significant,”
Psychological Science, vol. 22, no. 11, pp. 1359–1366, 2011.
DOI: 10.1177/0956797611417632.
[73] M. R. Munafò, B. A. Nosek, D. V. M. Bishop, K. S. Button,
C. D. Chambers, N. P. du Sert, et al., “A manifesto for
reproducible science,” Nature Human Behaviour, vol. 1,
art. 0021, 2017. DOI: 10.1038/s41562-016-0021.
[74] A. Zhang, Speech Recognition (Version 3.8), 2017.
[75] C. J. Hutto and E. Gilbert, “VADER: A parsimonious rule-based
model for sentiment analysis of social media text,” in
Proceedings of the Eighth International AAAI Conference on
Weblogs and Social Media (ICWSM-14), Ann Arbor, MI, USA,
The AAAI Press, 2014, pp. 216–225.
[76] D. Crystal and R. Quirk, Systems of prosodic and paralinguistic features in English, 39, Walter De Gruyter Inc., Berlin,
Germany, 1964.
[77] J. Pittam, Voice in social interaction: an interdisciplinary
approach, SAGE Publications, Thousand Oaks, California,
2020. DOI: 10.4135/9781483327105.
[78] Y. Jadoul, B. Thompson, and B. de Boer, “Introducing
Parselmouth: A Python interface to Praat,” Journal of Phonetics,
vol. 71, pp. 1–15, 2018. DOI: 10.1016/j.wocn.2018.07.001.
[79] P. Boersma and D. Weenink, “PRAAT, a system for doing
phonetics by computer,” Glot International, vol. 5,
pp. 341–345, 2001.
[80] H. M. Gray, K. Gray, and D. M. Wegner, “Dimensions of mind
perception,” Science, vol. 315, no. 5812, art. 619, 2007.
DOI: 10.1126/science.1134475.
[81] S. Cohen, T. Kamarck, and R. Mermelstein, “A global measure
of perceived stress,” Journal of Health and Social Behavior,
vol. 24, no. 4, pp. 385–396, 1983. DOI: 10.2307/2136404.
[82] G. Saucier, “Mini-markers: a brief version of Goldberg’s
unipolar Big-Five markers,” Journal of Personality
Assessment, vol. 63, no. 3, pp. 506–516, 1994.
DOI: 10.1207/s15327752jpa6303_8.
[83] S. Boll, “Suppression of acoustic noise in speech using
spectral subtraction,” IEEE Transactions on Acoustics,
Speech, and Signal Processing, vol. 27, no. 2, pp. 113–120,
1979. DOI: 10.1109/TASSP.1979.1163209.
[84] R. C. Arslan, M. P. Walther, and C. S. Tata, “formr: a study
framework allowing for automated feedback generation and
complex longitudinal experience-sampling studies using R,”
Behavior Research Methods, vol. 52, pp. 376–387, 2020.
DOI: 10.3758/s13428-019-01236-y.
[85] R. C. Arslan, C. S. Tata, and M. P. Walther, “formr: a study
framework allowing for automated feedback generation and
complex longitudinal experience sampling studies using R.”
(version v0.18.0), 2019. DOI: 10.5281/zenodo.3229668.
[86] S. Zhao, “Humanoid social robots as a medium of communication,” New Media & Society, vol. 8, no. 3, pp. 401–419,
2006. DOI: 10.1177/1461444806061951.
[87] J. F. Hoorn, “Theory of robot communication: I. The medium is
the communication partner,” arXiv preprint
arXiv:1812.04408, 2018.
[88] T. Abbas, G. Corpaccioli, V. J. Khan, U. Gadiraju, E. Barakova,
and P. Markopoulos, “How do people perceive privacy and
interaction quality while chatting with a crowd-operated
robot?” in Companion of the 2020 ACM/IEEE International
Conference on Human-Robot Interaction, Cambridge, UK:
Association for Computing Machinery, 2020, pp. 84–86.
DOI: 10.1145/3371382.3378332.
Tell me more! Assessing interactions with social robots from speech 157
[89] D. M. Hilty, K. Randhawa, M. M. Maheu, A. J. S. McKean,
R. Pantera, M. C. Mishkind, et al., “A review of telepresence,
virtual reality, and augmented reality applied to clinical
care,” Journal of Technology in Behavioral Science, vol. 5,
pp. 178–205, 2020. DOI: 10.1007/s41347-020-00126-x.
[90] E. van Wingerden, E. Barakova, T. Lourens, and P. S.
Sterkenburg, “Robot-mediated therapy to reduce worrying in
persons with visual and intellectual disabilities,” Journal of
Applied Research in Intellectual Disabilities, 2020,
DOI: 10.1111/jar.12801.
[91] M. Bunge, “A general black box theory,” Philosophy of
Science, vol. 30, no. 4, pp. 346–358, 1963. [Online].
Available: http://www.jstor.org/stable/186066.
[92] C. R. Berger and R. J. Calabrese, “Some explortations in initial
interaction and beyond: toward a developmental theory of
interpersonal communication,” Human Communication
Research, vol. 1, no. 2, pp. 99–112, 1975.
DOI: 10.1111/j.1468-2958.1975.tb00258.x.
[93] C. R. Berger and J. J. Bradac, Language and social knowledge:
uncertainty in interpersonal relations, London: Edward
Arnold, 1982.
[94] A. Waytz, J. Cacioppo, and N. Epley, “Who sees human?:
The stability and importance of individual differences in
anthropomorphism,” Perspectives on Psychological
Science, vol. 5, no. 3, pp. 219–232, 2010.
DOI: 10.1177/1745691610369336.
[95] K. Gray, L. Young, and A. Waytz, “Mind perception is the
essence of morality,” Psychological Inquiry, vol. 23, no. 2,
pp. 101–124, 2012. DOI: 10.1080/1047840X.2012.651387.
[96] D. M. Wegner, The Illusion of Conscious Will, Cambridge, MA:
MIT Press, 2002.
[97] E. Wiese, G. Metta, and A. Wykowska, “Robots as intentional
agents: using neuroscientific methods to make robots appear more social,” Frontiers in Psychology, vol. 8, art. 1663,
2017. DOI: 10.3389/fpsyg.2017.01663.
[98] V. J. Derlega, M. S. Harris, and A. L. Chaikin, “Selfdisclosure
reciprocity, liking and the deviant,” Journal of Experimental
Social Psychology, vol. 9, no. 4, pp. 277–284, 1973.
DOI: 10.1016/0022-1031(73)90065-6.
[99] N. Epley and A. Waytz, “Mind perception,” in Handbook of
Social Psychology, S. T. Fiske, D. T. Gilbert, and G. Lindzey
(Eds.), 5th edn, Hoboken, NJ, USA: John Wiley and Sons Ltd,
2010. DOI: 10.1002/9780470561119.socpsy001014.
[100] I. J. Firestone, “Reconciling verbal and nonverbal models of
dyadic communication,” Environmental Psychology and
Nonverbal Behavior, vol. 2, no. 1, pp. 30–44, 1977.
DOI: 10.1007/BF01127016.
[101] M. Argyle and M. Cook, Gaze and Mutual Gaze, Oxford,
England: Cambridge University Press, 1976.
[102] M. Argyle and J. Dean, “Eye-contact, distance and affiliation,”
Sociometry, vol. 28, no. 3, pp. 289–304, 1965.
DOI: 10.2307/2786027.
[103] M. L. Patterson, “Compensation in nonverbal immediacy
behaviors: a review,” Sociometry, vol. 36, no. 2, pp. 237–252,
1973. DOI: 10.2307/2786569.
[104] S. Chaiken, “Heuristic versus systematic information processing and the use of source versus message cues in persuasion,” Journal of Personality and Social Psychology, vol. 39,
no. 5, pp. 752–766, 1980. DOI: 10.1037/0022-3514.39.5.752.
[105] S. Chen, K. Duckworth, and S. Chaiken, “Motivated heuristic
and systematic processing,” Psychological Inquiry, vol. 10,
no. 1, pp. 44–49, 1999. DOI: 10.1207/s15327965pli1001_6.
[106] A. H. Eagly and S. Chaiken, The Psychology of Attitudes.
Orlando, FL, US: Harcourt Brace Jovanovich College
Publishers, 1993.
[107] L. A. Hosman, “The evaluational consequences of topic reciprocity and self-disclosure reciprocity,” Communication
Monographs, vol. 54, no. 4, pp. 420–435, 1987.
DOI: 10.1080/03637758709390242.
[108] T. Hirano, M. Shiomi, T. Iio, M. Kimoto, I. Tanev, K.
Shimohara, and N. Hagita, “How do communication cues
change impressions of human–robot touch interaction?”
International Journal of Social Robotics, vol. 10, no. 1,
pp. 21–31, 2018. DOI: 10.1007/s12369-017-0425-8.
[109] M. Iwasaki, J. Zhou, M. Ikeda, Y. Koike, Y. Onishi, T.
Kawamura, and H. Nakanishi, ““That robot stared back at
me!”: Demonstrating perceptual ability is key to successful
human–robot interactions,” Frontiers in Robotics and AI,
vol. 6, 2019. DOI: 10.3389/frobt.2019.00085.
[110] C. J. A. M. Willemse and J. B. F. van Erp, “Social touch in
human–robot interaction: robot-initiated touches can induce
positive responses without extensive prior bonding,”
International Journal of Social Robotics, vol. 11, pp. 285–304,
2019. DOI: 10.1007/s12369-018-0500-9.
[111] D. J. Bem, “Self-perception: an alternative interpretation of
cognitive dissonance phenomena,” Psychological
Review, vol. 74, no. 3, pp. 183–200, 1967.
DOI: 10.1037/h0024835.
[112] D. J. Bem, “Self-perception theory,” Advances in
Experimental Social Psychology, vol. 6, pp. 1–62, 1972.
DOI: 10.1016/S0065-2601(08)60024-6.
[113] D. Kahneman and A. Tversky, “Subjective probability: a judgment of representativeness,” Cognitive Psychology, vol. 3, no. 3,
pp. 430–454, 1972. DOI: 10.1016/0010-0285(72)90016-3.
[114] D. L. Schacter, “The seven sins of memory: insights from
psychology and cognitive neuroscience,” American
Psychologist, vol. 54, no. 3, pp. 182–203, 1999.
DOI: 10.1037/0003-066X.54.3.182.
[115] M. J. Mahoney, “Publication prejudices: an experimental
study of confirmatory bias in the peer review system,”
Cognitive Therapy and Research, vol. 1, no. 2, pp. 161–175,
1977. DOI: 10.1007/BF01173636.
[116] N. Akalin and H. Köse, “Emotion recognition in valencearousal scale by using physiological signals,” in 2018 26th
Signal Processing and Communications Applications
Conference (SIU), Izmir, Turkey: IEEE, 2018, pp. 1–4.
DOI: 10.1109/SIU.2018. 8404632.
[117] C. Breazeal and L. Aryananda, “Recognition of affective
communicative intent in robot-directed speech,”
Autonomous Robots, vol. 12, no. 1, pp. 83–104, 2002.
DOI: 10.1023/A:1013215010749.
[118] K. Fischer, O. Niebuhr, L. C. Jensen, and L. Bodenhagen, “Speech
melody matters — How robots profit from using charismatic
speech,” ACM Transactions on Human-Robot Interaction (THRI),
vol. 9, no. 1, pp. 1–21, 2019. DOI: 10.1145/3344274.
[119] E. S. Kim, D. Leyzberg, K. M. Tsui, and B. Scassellati, “How
people talk when teaching a robot,” in Proceedings of the 4th
ACM/IEEE International Conference on Human-Robot
158 Guy Laban et al.
Interaction, Association for Computing Machinery (ACM),
La Jolla, California, USA, 2009, pp. 23–30.
DOI: 10.1145/1514095.1514102.
[120] A. L. Robinson-Mosher and B. Scassellati, “Prosody recognition in male infant-directed speech,” 2004 IEEE/RSJ
International Conference on Intelligent Robots and Systems
(IROS) (IEEE Cat. No. 04CH37566), IEEE, Sendai, Japan, vol. 3,
pp. 2209–2214, 2004. DOI: 10.1109/IROS.2004.1389737.
[121] G. Skantze, C. Oertel, and A. Hjalmarsson, “User feedback in
human–robot interaction: prosody, gaze and timing,”
Proceedings of the Annual Conference of the International
Speech Communication Association, Lyon, France: INTERSPEECH, pp. 1901–1905, 2013.
[122] J. O. Greene, “A cognitive approach to human communication: an action assembly theory,” Communication
Monographs, vol. 51, no. 4, pp. 289–306, 1984.
DOI: 10.1080/03637758409390203.
[123] J. O. Greene, “Production of messages in pursuit of multiple
social goals: action assembly theory contributions to the
study of cognitive encoding processes,” Annals of the
International Communication Association, vol. 18, no. 1,
pp. 26–53, 1995. DOI: 10.1080/23808985.1995.11678906.