Nothing Special   »   [go: up one dir, main page]

Skip to main content

SYSTEMATIC REVIEW article

Front. Robot. AI, 11 May 2021
Sec. Human-Robot Interaction

A Systematic Review of Robotic Rehabilitation for Cognitive Training

  • 1Department of Mechanical, Aerospace, and Biomedical Engineering, University of Tennessee, Knoxville, Knoxville, TN, United States
  • 2School of Nursing, MGH Institute of Health Professions, Boston, MA, United States

A large and increasing number of people around the world experience cognitive disability. Rehabilitation robotics has provided promising training and assistance approaches to mitigate cognitive deficits. In this article, we carried out a systematic review on recent developments in robot-assisted cognitive training. We included 99 articles in this work and described their applications, enabling technologies, experiments, and products. We also conducted a meta analysis on the articles that evaluated robot-assisted cognitive training protocol with primary end users (i.e., people with cognitive disability). We identified major limitations in current robotics rehabilitation for cognitive training, including the small sample size, non-standard measurement of training and uncontrollable factors. There are still multifaceted challenges in this field, including ethical issues, user-centered (or stakeholder-centered) design, the reliability, trust, and cost-effectiveness, personalization of the robot-assisted cognitive training system. Future research shall also take into consideration human-robot collaboration and social cognition to facilitate a natural human-robot interaction.

1. Introduction

It is estimated that ~15% of the world's population, over a billion people, experience some form of disability and a large proportion of this group specifically experience cognitive disability (WHO, 2011). The number of people with disabilities is increasing not only because of the growing aging population who have a higher risk of disability but also due to the global increase in chronic health conditions (Hajat and Stein, 2018). Individuals with cognitive disability, such as Alzheimer's disease (AD) or Autism spectrum disorder (ASD), may have a substantial limitation in their capacity for functional mental tasks, including conceptualizing, planning, sequencing thoughts and actions, remembering, interpreting subtle social cues, and manipulating numbers and symbols (LoPresti et al., 2008). This vulnerable population is usually associated with significant distress or disability in their social, occupational, or other important activities.

With recent advancements of robotics and information and communication technologies (ICTs), rehabilitation robots hold promise in augmenting human healthcare and in aiding exercise and therapy for people with cognitive disabilities. As an augmentation of human caregivers with respect to the substantial healthcare labor shortage and the high burden of caregiving, robots may provide care with high repeatability and without any complaints and fatigue (Taheri et al., 2015b). In a meta analysis comparing how people interacted with physical robots and virtual agents, Li (2015) showed that physically present robots were found to be more persuasive, perceived more positively, and result in better user performance compared to virtual agents. Furthermore, robots can facilitate social interaction, communication and positive mood to improve the performance and effectiveness of cognitive training (Siciliano and Khatib, 2016). For example, a recent study (Pino et al., 2020) showed that older adults with mild cognitive impairment (MCI) that received memory training through the humanoid social robot (NAO) achieved more visual gaze, less depression, and better therapeutic behavior. Physically embodied robots hold promise as accessible, effective tools for cognitive training and assistance in future.

There have been a few literature reviews on physical rehabilitation (Bertani et al., 2017; Kim et al., 2017; Morone et al., 2017; Veerbeek et al., 2017), or cognitive rehabilitation for specific user populations, such as children with ASD (Pennisi et al., 2016) and older adults (Mewborn et al., 2017). To the authors' best knowledge, this article presents the first systematic review that focuses on robotic rehabilitation for cognitive training. We present applications, enabling technologies, and products of robotics rehabilitation based on research papers research papers focusing on cognitive training. We also discuss several challenges to the development of robots for cognitive training and present future research directions.

2. Methods

2.1. Search Strategy

We conducted a systematic review in the datasets of Google Scholar, Crossref, PubMed, Scopus, and Web of Science using the key words (“robot” OR “robots” OR “robotics” OR “robotic”) AND (“cognitive training” OR “cognitive rehabilitation” OR “cognitive therapy” OR “cognitive recovery” OR “cognitive restore”). The search was limited to the articles published between 2015 and December 14, 2020. The search in Google Scholar yielded 5,630 articles. Google Scholar ranks sources according to relevance, which takes into account the full text of each source as well as the source's author, the publication in which the source appeared and how often it has been cited in scholarly literature (University of Minnesota Libraries, 2021). We screened the titles of the first 500 articles and excluded the remaining results due to their low relevance. After the analysis of abstracts and written languages, 328 articles were further excluded due to duplication (i.e., exact copy of two works), non-English language, and/or not pertaining to the research topic. Then 172 full articles were reviewed for eligibility. The articles that were aimed for physical rehabilitation or review articles were excluded. Finally, 80 eligible articles were included for further analysis. As illustrated in the PRISMA flow diagram (see Figure 1), with the keywords we initially found 200, 31, 106, and 50 articles in the datasets of Crossref, PubMed, Scopus and Web of Science for screening, respectively. These articles were combined with the 80 additional articles from Google Scholar for further screening. After the analysis of titles, abstracts, written languages and types of article, 238 articles were excluded due to duplication, non-English language, non-eligible article types (e.g., book chapter, book, dataset, report, reference entry, editorial and systematic review), and/or not pertaining to the research topic. Then 229 full articles were reviewed for eligibility. The articles that were aimed for physical rehabilitation or review articles were excluded. Finally, 99 eligible articles were included in this systematic review, including journal articles and conference papers presented at conference, symposium and workshop. These papers included the contents of applications, user population, supporting technologies, experimental studies, and/or robot product(s). Moreover, 53 articles that included experimental study of robot-assisted cognitive training with primary end users (i.e., people with cognitive disability) were identified for further meta-analysis.

FIGURE 1
www.frontiersin.org

Figure 1. PRISMA flow diagram illustrating the exclusion criteria and stages of the systematic review.

The literature include cognitive training robots in the forms of companion robots, social robots, assistive robots, social assisted robots, or service robots, which are collectively referred to as “cognitive training robots” in this systematic review. The literature employed different terminologies for cognitive training by rehabilitation robots, such as robot-enhanced therapy (David et al., 2018; Richardson et al., 2018), robot-assisted intervention (Scassellati et al., 2018), robot-assisted treatment (Taheri et al., 2015a), robot-assisted training (Tsiakas et al., 2018), robot-assisted therapy (Sandygulova et al., 2019), robot-mediated therapy (Begum et al., 2015; Huskens et al., 2015). Here, we do not distinguish between these different terms and instead adopt the term of “robot-assisted training” to represent all these different terms.

3. Results

3.1. Applications

The studies on robot-assisted cognitive training are categorized in terms of their applications and end users in Table 1. To date, the most researched application (36 out of 98 articles, as shown in Table 1) of robots in cognitive training is to improve individual social communication skills, which may include joint attention skills, imitation skills, turn-taking skills and other social interaction skills. For example, Kajopoulos et al. (2015) designed a robot-assisted training protocol based on response to joint attention for children with ASD. The training protocol used a modified attention cueing paradigm, where the robot's head direction cued children's spatial attention to a stimulus presented on one side of the robot. The children were engaged in a game that could be completed only through following the robot's head direction. To aid in the training of imitation skills in children with ASD, Taheri et al. (2020) proposed a reciprocal gross imitation human–robot interaction platform, in which ASD children are expected to imitate the robot body gestures, including a combination of arms, feet, neck, and torso movements. David et al. (2020) developed a robot-enhanced intervention on turn-taking abilities in ASD children. In their protocol, the robot provided instruction (e.g., “Now is my turn”) to the child, checked if the child moved the picture as instructed, and provided feedback (e.g., “Good job”) to the child if the child respected turns by staying with his or her hands still, without interrupting the robot. Robot-assisted cognitive training showed increased cognitive capabilities for people with limited social capabilities, such as children with ASD (Huijnen et al., 2016; Esteban et al., 2017; Marino et al., 2019) and people with dementia (Sung et al., 2015; Yu et al., 2015; Otaki and Otake, 2017). Due to cognitive impairment, individuals with dementia may also show deficits in social functioning, such as social withdrawal (Havins et al., 2012; Dickerson, 2015). In the pilot study by Sung et al. (2015) about robot-assisted therapy using socially assistive pet robot (PARO), institutionalized older adults showed significantly increased communication and interaction skills and activity participation after receiving 4-week robot-assisted therapy. Another robotic application is to provide intervention to enhance people's impaired cognitive function, such as memory (Paletta et al., 2018), attention (Lins et al., 2019) and concentration (Tleubayev et al., 2019), or reduce their negative psychophysiological feelings, such as stress (Aminuddin et al., 2016) and anxiety (Ab Aziz et al., 2015). Additionally, a few studies adopted the robots to facilitate learning and educational activities for people with cognitive disabilities, such as children with dyslexia (Andruseac et al., 2015) or Traumatic Brain Injury (TBI) (Barco Martelo and Fosch Villaronga, 2017).

TABLE 1
www.frontiersin.org

Table 1. Type of robot-assisted cognitive training and end-user population.

3.2. Enabling Technologies

Recent development in robotics, ICTs, multimodal human-robot interaction, and artificial intelligence leads to significant process in robot-assisted cognitive training and rehabilitation. This section presents a summary on a few important enabling technologies that foster the advancement of robotic rehabilitation for cognitive training, including multimodal perception, multimodal feedback, gamification, virtual and augmented reality, and artificial intelligence.

3.2.1. Design of Physical Appearance

A robot can have a human-like (Miskam et al., 2015; Peleka et al., 2018; Taheri et al., 2019), animal-like (Cao et al., 2015; Sung et al., 2015), or unfamiliar appearance (Scassellati et al., 2018). Besides the appearance, the size, softness, and comfort of the robot can also have an impact on users' perception, affection, cognitive load, and gaze following during interaction (Kohori et al., 2017), and thus the effectiveness of cognitive training. It remains unclear how users' perception is specifically affected by the robot's appearance. On the one hand, the human-like appearance was indicated to significantly positively affect users' perception of anthropomorphism, animacy, likeability, and intelligence toward robots, compared to a snowman-like appearance (Haring et al., 2016). On the other hand, increasing human-like appearance was found not to necessarily increase performance in human-robot interaction. For example, in the survey of expectation about the robots' appearance in robot-assisted ASD therapy, zoomorphic robots were indicated to be less ethically problematic than robots that looked too much like humans (Coeckelbergh et al., 2016). Some of their participants (i.e., parents, therapists, or teachers of children with ASD) worried about the possibility that the robot is perceived by the child as a friend, or that the robot looks too human-like. The relation between robots' human-like appearance and people's reaction to them may relate to the uncanny valley theory (Mori et al., 2012), which describes that people's familiarity (or affinity) with robots increases as the robots appear more human-like but when the robots are almost human, people's response would shift to revulsion. Tung (2016) observed an uncanny valley between the degree of anthropomorphism and children's attitudes toward the robots. In the review paper to study factors affecting social robot acceptability in older adults including people with dementia or cognitive impairment, Whelan et al. (2018) found that there is a lack of consensus regarding to the optimal appearance of social robots and that the uncanny valley concept varies between individuals and groups.

3.2.2. Multimodal Sensing

Having a good understanding of a user's cognitive state, affective state and surrounding environment, which is termed as multimodal perception, is a prerequisite step for robots to provide cognitive training. Usually the concept of multimodal perception involves two stages, multimodal sensing and perception. Further details on these two techniques in previous publications are individually discussed in the following. Various sensors have been adopted to facilitate a robot to achieve multimodal sensing, based on system requirements, end-user population, cost-effectiveness, etc. Among different sensing technologies, visual and auditory sensing are the most popular modalities. We summarize the multiple modalities for sensing in the following aspects.

1. Visual sensing. During human-robot interaction, visual sensors are a very popular, useful and accessible channel for perception. The advancement of technologies, such as manufacturing and ICTs, enabled researchers to integrate small, high-resolution and affordable cameras into their rehabilitation robotic system. Some studies placed cameras in the environment along with the robot (Melo et al., 2019). Other studies included cameras in the robotic mechanical system. For example, in the social robot Pepper, there were 2D and 3D cameras attached to the head (Paletta et al., 2018). With computational approaches, such as computer vision, the robots analyzed the video/images from the cameras and recognized users' critical states, such as their environment, facial expression, body movements, and even emotion and attention (Paletta et al., 2018; Peleka et al., 2018; Rudovic et al., 2018; Mois et al., 2020), which led to a better perception of users (Johal et al., 2015).

2. Auditory sensing. Another popular modality adopted during robot-assisted cognitive training is auditory sensing. Researchers analyzed users' auditory signals in terms of the lexical field, tone and/or volume of voice for speech recognition, emotion detection, and speaker recognition in robots (Paletta et al., 2018; Peleka et al., 2018; Rudovic et al., 2018). Due to the natural and intuitive nature for users behind the auditory sensing, some end users, such as older adults preferred this sensing channel to the touch input during interaction with the robot (Zsiga et al., 2018).

3. Physiological sensing. Besides visual and auditory sensing, physiological modalities have been incorporated into the robotic system, in order to have a better understanding of users' states (e.g., affective states). Previous studies show that human-robot interaction may be enhanced using physiological signals, such as heart rate, blood pressure, breathing rate, and eye movement (Sonntag, 2015; Lopez-Samaniego and Garcia-Zapirain, 2016; Ozcana et al., 2016; Ahn et al., 2017; Alimardani and Hiraki, 2017). For example, Rudovic et al. (2018) employed wearable sensors to detect children's heart-rate, skin-conductance and body temperature during their robot-based therapy for children with ASD, to estimate children's levels of affective states and engagement. When studying robot-assisted training for children with ASD, Nunez et al. (2015) used a wearable device of electromyography (EMG) sensors to detect smiles from children's face.

4. Neural sensing. The inclusion of brain-imaging sensors provides capabilities to measure and/or monitor a user's brain activity and to understand the user's mental states (Ali et al., 2019). This was especially useful when considering the user's cognitive states, such as level of attention and task engagement (Alimardani and Hiraki, 2017; Tsiakas et al., 2018). Neural sensing may also be meaningful to users who have difficulty in expressing their intention and feeling because of their physical and/or cognitive limitations, such as the older adults with Alzheimer's disease. Currently, among all candidates of brain-imaging sensors, EEG and functional near-infrared spectroscopy (fNIRS) are two popular modalities due to their advantages of non-invasiveness, portability, and cost-effectiveness. For example, Lins et al. (2019) developed a robot-assisted therapy to stimulate the attention level of children with cerebral palsy and applied electroencephalogram (EEG) sensors to measure children's attention level during the therapy.

3.2.3. Multimodal Feedback

After perceiving its user and environment, a robot shall entail multimodal feedback to interact (e.g., display its behaviors and feedback) with its user in a comfortable, readable, acceptable, and effective way (Melo et al., 2019). Multimodal feedback is particularly meaningful when the end users are unfamiliar with technologies or are limited in cognitive capabilities, such as older adults with dementia. Examples of feedback include voice, video, gesture, and physical appearance, all of which can affect users' perception of the robot during their interaction and thus the effectiveness of cognitive training/rehabilitation (Ab Aziz et al., 2015; Rabbitt et al., 2015). The following list shows popular modalities for robotic feedback during interaction with human users.

1. Visual feedback. One of the most widely-used overt feedback modalities is visual feedback or graphical user interface (GUI), displaying two-dimensional information. During robot-assisted cognitive training, visual feedback was delivered through an additional computer screen (Salvador et al., 2016; Taheri et al., 2019; Mois et al., 2020), or a touchscreen embodied in the robot (Paletta et al., 2018; Peleka et al., 2018). Some principles and/or issues on the design of GUI for robot-assisted training were suggested in previous studies. For example, a few studies recommended a larger screen and a simpler interface associated with each function choice for GUI to better facilitate visual feedback during cognitive training (Ahn et al., 2017).

2. Auditory feedback. Another widely used modality for feedback or interaction is auditory output, like speech during human-human interaction. This intuitive auditory communication can reduce the unfamiliarity with robotic technologies and increase the usability of system in the vulnerable population, such as the elderly population (Zsiga et al., 2018). Among previous studies, this feedback was delivered in one or combined form of beep (Nault et al., 2020), speech (Ab Aziz et al., 2015; Miskam et al., 2015; Peleka et al., 2018; Taheri et al., 2019), and music (Nunez et al., 2015). With auditory output, the robot provided daily communication and medication reminder, instructed cognitive training, and made emergent warning (e.g., short of battery) to the user (Orejana et al., 2015).

3. Non-verbal feedback. We refer to non-verbal feedback as all non-verbal communication cues by a robotic body, such as hand gestures and other body movements (Miskam et al., 2015; Taheri et al., 2019), eye gaze (Taheri et al., 2019), eye colors (Miskam et al., 2015; Taheri et al., 2019), and facial expression. Animation, similar to verbal language, makes a significant contribution to improving robot-assisted cognitive training. Moreover, the robot's animation can particularly introduce social interaction to a user, which is meaningful to individuals with impairments in social interaction skills.

4. Haptic feedback. Haptic feedback, by simulating the sense of touch and motion, may play important role during robot-assisted cognitive training due to the importance of touch in everyday life (Beckerle et al., 2017; Cangelosi and Invitto, 2017). Tactile feedback is one type of haptic feedback. During robot-assisted cognitive training, haptic feedback can also be introduced using vibration via wearable devices. For example, Nault et al. (2020) developed a socially assistive robot-facilitated memory game elaborated with audio and/or haptic feedback for older adults with cognitive decline. Although there was no significant difference in participants' game accuracy, preference, and performance in their system pilot study, the results provided insight into future improvements, such as increasing the strength of haptic feedback to increase the ease of being perceived and make the system more engaging. One notable robot, Paro, with the combination of soft, plush surface and additional encouraging haptic feedback (e.g., small vibration) creates a soothing presence. In a pilot study for institutionalized older adults by Sung et al. (2015), the communication and social skills of participants were improved by the robot-assisted therapy using Paro.

3.2.4. Gamification

Recently, game technology is becoming a popular way to motivate, engage and appeal to users in cognitive tasks, since traditional cognitive tasks are usually effortful, frustrating, repetitive, and disengaging. Serious games and brain training games are a growing research field for cognitive training (Heins et al., 2017). Integration between gaming and robotic technologies has attracted increasing amount of interest in research and application, to further enhance users' engagement in cognitive training. A few types of relationships between robots and games have been developed in the literature. For example, a robot can lead or accompany users through the game for cognitive training by providing instructions on how to perform the task (Ioannou et al., 2015; Chu et al., 2017; Tsardoulias et al., 2017; Scassellati et al., 2018; Sandygulova et al., 2019; Taheri et al., 2019; Tleubayev et al., 2019; Nault et al., 2020) or playing a role of an active player in the game (Tariq et al., 2016; Melo et al., 2019). Additionally, a robot can provide various types of feedback (see details in section 3.2.3) to encourage users to engage in the game (Taheri et al., 2015a; Lopez-Samaniego and Garcia-Zapirain, 2016). Often, games associated with cognitive training can be integrated into the robotic systems through a GUI (Ahn et al., 2017; Paletta et al., 2018; Peleka et al., 2018).

3.2.5. Virtual and Augmented Reality

Combining robot-assisted cognitive training with virtual reality (VR) and/or augmented reality (AR) techniques offers a cost-effective and efficient alternative to traditional training settings. The incorporation of VR/AR allows for replication of the tasks and environments in a more convenient and affordable way. Researchers also explored robotic cognitive training using mixed reality technology in cognitive training. For example, Sonntag (2015) presented an intelligent cognitive enhancement platform for people with dementia, where a mixed reality glass was used to deliver the storyboard (e.g., serious game for active memory training) to the user and a NAO robot served as a cognitive assistant for “daily routine.” Bozgeyikli et al. (2015) used virtual reality in a vocational rehabilitation system, which included six different modules, such as money management in a virtual grocery store, to provide vocational training for persons with ASD and TBI.

3.2.6. Artificial Intelligence

Artificial Intelligence plays a significant role in the field of robot-assisted cognitive training/rehabilitation, including applications in multimodal perception and feedback, personalization, and adaptability (Ab Aziz et al., 2015; Rudovic et al., 2018). Given multimodal sensing, a successful multimodal perception further requires robots to integrate signals across multiple modalities of input sensors. To date, a great progress has been made thanks to the advancement of machine learning and deep learning. Multi-modal signals enable the robot with a good interpretation and understanding of its users, including their needs, intention, emotions, and surrounding environment (Paletta et al., 2018). Rudovic et al. (2018) implemented deep learning in a robot for ASD therapy to automatically estimate children's valence, arousal and engagement levels. Javed et al. (2018) utilized multimodal perception, including the analyses of children's motion, speech, and facial expression, to estimate children's emotional states. Using multiple feedback modalities may overload users with redundant information, increase task completion time, and reduce the efficiency of cognitive training (Taranović et al., 2018c). Additionally, users may favor certain modalities over others due to personal preference or cognitive disability. Taranović et al. (2018c) designed an experiment of adaptive modality selection (AMS) in robot-assisted sequential memory exercises and applied artificial intelligence to learn the strategy that selects the appropriate combination and amount of feedback modalities tailored to different situations (e.g., environments and users). An appropriate strategy is crucial to successful long-term robot-assisted cognitive intervention. Specifically, reinforcement learning, an area of machine learning, is a promising approach to adapt and personalize the intervention to each individual user as well as to optimize the performance of robot-assisted cognitive training, due to its capability of allowing a robot to learn from its experience of interaction with users (Sutton and Barto, 2018). For example, Tsiakas et al. (2016) used interactive reinforcement learning methods to facilitate the adaptive robot-assisted therapy, that is, adapt the task difficulty level and task duration to users with different skill levels (e.g., expert or novice user), in the context that users need to perform a set of cognitive or physical training tasks. Javed et al. (2018) developed a Hidden Markov model (HMM) in their adaptive framework for child-robot interaction, aiming to enable a child with ASD to engage in robot-assisted ASD therapy over long term. In their HMM, the states took into consideration a child's emotional state or mood, and the actions were the robot's behaviors or other audiovisual feedback. Clabaugh et al. (2019) utilized reinforcement learning to personalize instruction challenge levels and robot feedback based on each child's unique learning patterns for long-term in-home robot interventions. Although reinforcement learning may suffer the problem of sample inefficiency, the slowness of reinforcement learning can be overcome using techniques, such as episodic memory and meta-learning (Botvinick et al., 2019).

3.3. Experimental Studies

Many experimental studies have been conducted to evaluate the important properties of robotic rehabilitation, such as feasibility, safety, usability, performance, etc. On the one hand, exploratory studies including surveys and interviews with users (e.g., patients, caregivers, and therapists) have been conducted to inform the next stage of study (Rabbitt et al., 2015; Coeckelbergh et al., 2016; Salichs et al., 2016; Darragh et al., 2017; Kohori et al., 2017; Korchut et al., 2017; Law et al., 2019b). On the other hand, researchers have conducted experimental studies to verify and/or validate robot-assisted cognitive training systems. Table 2 shows a meta analysis for experimental studies, where the robot-assisted cognitive training was provided to primary end users (i.e., persons with cognitive disabilities). Up to date, majority of the experimental studies were conducted in a controlled lab setting, and only a few studies were conducted in an environment simulating daily activities in real world (Scassellati et al., 2018).

TABLE 2
www.frontiersin.org

Table 2. Meta analysis on end-user experiments of robot-assisted cognitive training.

3.3.1. Study Design

Most experimental studies included three phases: pre-training assessment (i.e., baseline assessment), robot-assisted cognitive training, and post-training assessment (Kajopoulos et al., 2015; Kim et al., 2015; Sung et al., 2015; Yu et al., 2015; Alemi et al., 2016; Taheri et al., 2016, 2019; van den Heuvel et al., 2017; Scassellati et al., 2018; Marino et al., 2019). The effectiveness of robot-assisted training was evaluated by the comparison of pre- and post-training assessments using machine learning or statistical methods (Kim et al., 2015; Yu et al., 2015; Scassellati et al., 2018; Marino et al., 2019; Taheri et al., 2019).

Most studies adopted the group-based design where participants were randomly assigned to control or intervention groups (Kim et al., 2015; Sung et al., 2015; Yu et al., 2015; Marino et al., 2019). Some researchers employed single-case designs (or single-subject designs) to investigate the impact of social robots on cognitive training (Ioannou et al., 2015; Taheri et al., 2015a; David et al., 2018). For example, Ioannou et al. (2015) conducted the single-case study to explore the potential role of co-therapist of humanoid social robot, NAO, during autism therapy session with one child with ASD. In their study, there are four intervention sessions, and one follow-up, post-intervention therapy session to examine the effectiveness of the therapy with NAO.

Sample sizes vary dramatically in the literature (Ioannou et al., 2015; Kajopoulos et al., 2015; Kim et al., 2015; Sung et al., 2015; Chu et al., 2017; Khosla et al., 2017; Rudovic et al., 2018), where some studies were conducted with hundreds of participants whereas some studies included only a few participants. Challenges to recruitment included accessibility of participants and their caregivers, participants' disability, and ethical issues (e.g., privacy).

In terms of the intensity and duration of robot-assisted cognitive training, due to the variety of applications and end users, there was also a great variation in the total number of training sessions as well as the session duration. For example, with respect to one single training session, it took from about 20 min (Shukla et al., 2015; Tleubayev et al., 2019) to 90 min (Kim et al., 2015). Corresponding to the total cognitive training period, it varied from a few days (Bharatharaj et al., 2017) to more than 5 years (Chu et al., 2017).

3.3.2. Evaluation

Researchers employed subjective and/or objective evaluation metrics to evaluate the performance of robots in cognitive training. Subjective measurement include qualitative observation, interviews and questionnaires. Objective measurements evaluate the performance from a behavioral or neurophysiological level.

1. Observation. During the experiments or recorded video, the experimenters or professional therapists observed and evaluated participants' behaviors, such as affective feelings, eye contact, communication, and other related interactions, based on their knowledge and experience (Begum et al., 2015; Conti et al., 2015; Costescu et al., 2015; Ioannou et al., 2015; Shukla et al., 2015; Taheri et al., 2015a, 2019; Yu et al., 2015; Tariq et al., 2016; Wong and Zhong, 2016; Yun et al., 2016; Zheng et al., 2016; Abdollahi et al., 2017; Chu et al., 2017; Garcia-Sanjuan et al., 2017; Khosla et al., 2017; Rudovic et al., 2017; David et al., 2018; Marino et al., 2019; Sandygulova et al., 2019; Tleubayev et al., 2019). This measurement was a very practical, dominant metric during the study of children with ASD. With the development of ICTs, some studies also applied customized software (instead of human effort) to evaluate user's behaviors, such as smiles and visual attention (Pino et al., 2020).

2. Interview. Interviews were conducted with the primary users (i.e., patients), their caregivers (e.g., parents and other family caregivers), and therapists, to learn about users' opinion and experience, and the performance of the robot-assisted cognitive training (Yu et al., 2015; Bharatharaj et al., 2017; Darragh et al., 2017; Paletta et al., 2018; Sandygulova et al., 2019; Taheri et al., 2019; Tleubayev et al., 2019). As stated in the book on user experience (UX) lifecycle by Hartson and Pyla (2018), user interview is a useful, practical technique to understand users' needs, design solutions, and evaluate UX, all of which are basic fundamental activities in UX lifecycle. Specifically, interviews can be applied to extract requirements of people with cognitive disability and/or their caregivers, to create the human-robot interaction design concepts, and to verify and refine human-robot interaction design for cognitive training. For example, in the case studies by Orejana et al. (2015), older adults with chronic health conditions in a rural community used a healthcare robot (iRobi) in their homes for at least 3 months. Then participants were interviewed to learn personal accounts of participants's experience. Through the interview, the authors found that more familiar games may be easier for older people to relate to and therefore increase users' confidence and that a larger screen would make the functions easier to see and use. The interview also revealed that older people sometimes have less dexterity so making the touchscreen less sensitive to long presses may remove accidental triggering of functions.

3. Questionnaire. Most studies utilized questionnaires to evaluate the performance of robot-assisted cognitive training. Researchers adopted questionnaire(s) based on their targeted performance, such as targeted user groups (e.g., patients, caregivers, or therapists), targeted cognitive capabilities (e.g., memory or anxiety), and research goals (e.g., users' perception of robot or effectiveness of robot). A few studies designed their own questionnaires according to their study (Tariq et al., 2016; Abdollahi et al., 2017; Ahn et al., 2017; Bharatharaj et al., 2017; Khosla et al., 2017; van den Heuvel et al., 2017; Scassellati et al., 2018; Lins et al., 2019; Tokunaga et al., 2019). Table 3 shows a list of common questionnaires in the literature.

TABLE 3
www.frontiersin.org

Table 3. Questionnaires used to evaluate performance of robot-assisted cognitive training.

4. Behavioral measurement. From a behavioral perspective, researchers measured the number of correct/incorrect responses, response time, and/or time to complete the activity by participants to evaluate the performance of robot-assisted training (Bozgeyikli et al., 2015; Costescu et al., 2015; Ioannou et al., 2015; Kajopoulos et al., 2015; Salvador et al., 2016; Shukla et al., 2017; Lins et al., 2019; Nault et al., 2020).

5. Neurophysiological measurement. The advancement of brain-imaging technologies and deep learning enables researchers to assess the impact of cognitive training on cognitive capabilities from a neurophysiological perspective, using brain-imaging tools, such as EEG, fNIRS or functional magnetic resonance imaging (fMRI) (Ansado et al., 2020). Researchers also applied such tools to detect changes in the brain associated with participants' cognitive capability as metrics to evaluate the performance of robots in cognitive training (Kim et al., 2015; Alimardani and Hiraki, 2017).

3.4. Robot Products

The development of technologies, such as manufacturing and ICTs, has led to the generation of mass-product robots for research, education and therapeutic applications (Wood et al., 2017; Pandey and Gelin, 2018). Particularly in the field of cognitive training/rehabilitation, the developed robots are featured with capabilities, such as the aforementioned multimodal perception and multimodal feedback to support the human-robot interaction during cognitive training. Table 4 shows commonly used robot products as well as the important features to enable these robots to assist cognitive training. Their specific applications in cognitive training among previous studies, for example, assisting the intervention for memory and social communication skills, are listed in Table 1.

TABLE 4
www.frontiersin.org

Table 4. Features of common robot products for cognitive training.

4. Discussions

4.1. Limitations

4.1.1. Sample Size

Probably, the most common challenge faced by researchers of cognitive training is a small size of participants. This imposes the generalization and reliability of experimental results in question. The limitation of small sample size was caused by the small number of participants or the limited number of available robots for experiments (Kajopoulos et al., 2015; Shukla et al., 2015; Zheng et al., 2015; Tariq et al., 2016; Bharatharaj et al., 2017; Darragh et al., 2017; Garcia-Sanjuan et al., 2017; Tsardoulias et al., 2017; Lins et al., 2019; Marino et al., 2019; Tleubayev et al., 2019). Typically, a research lab has only a few robots. This would be particularly challenging to experimental studies that require multiple sessions for each individual user. In this case, within the same study period, these studies have to control the number of participants to a small number. The recruitment of participants was influenced by the accessibility to participants during the whole study and sometimes the problems associated with their caregivers (Alemi et al., 2016; Taheri et al., 2019). For example, in the study (Alemi et al., 2016) of exploring effect of utilizing a robot NAO as a therapy-assistive tool to deal with pediatric distress among children with cancer. In terms of the small sample size, researchers mentioned that due to the novelty of their project and scant number of systematical psychological interventions for patients with cancer or other refractory illness in Iranian hospitals, it was difficult to persuade children's parents to join their study. Also, they mentioned that it was hard for parents to bring their kids to intervention sessions on a regular basis. Moreover, the issue of small sample size also means that the participants in some studies were not general and representative, in terms of factors, such as their severity of cognitive disability, their age and gender (Begum et al., 2015; Yu et al., 2015; Chu et al., 2017).

4.1.2. Measurement of Training Effectiveness

Another impeding factor in studies of robot-assisted cognitive training is the evaluation of its training effectiveness, which can relate to choosing tools for specific evaluation metrics, identifying relevant evaluation metrics, or designing experiments to facilitate evaluation. In terms of the evaluation metrics, many studies adopted subjective evaluations, which could be biased and inaccurate. To the authors' best knowledge, there is no standardized questionnaire to evaluate robot-assisted cognitive training. As shown in Table 3, multiple different questionnaires were applied to evaluate the same target (e.g., robot acceptance), which makes it difficult to compare the performance of robot-assisted cognitive training between different studies (Bharatharaj et al., 2017). Often, assessment metrics focus on the impact on the specifically trained cognitive capability, ignoring the potential transfer to other cognitive skills (Zheng et al., 2016) and the long-term performance (Richardson et al., 2018). Moreover, evaluations were frequently conducted for the robot-assisted cognitive training in controlled laboratory settings. The real-world environments are usually noisy and dynamic, which brings greater challenges for a reliable, robust user-robot interaction and a good user experience of the robot (Salem et al., 2015; Trovato et al., 2015).

Additionally, the effectiveness of robot-assisted cognitive training may be impacted by users' perceived interaction with the robot. On the one hand, some studies (Lopez-Samaniego and Garcia-Zapirain, 2016; Pereira et al., 2019; Mois et al., 2020) have evaluated the acceptance, satisfaction and perception of robots for cognitive training. On the other hand, many studies (Kim et al., 2015; Shukla et al., 2015; Demetriadis et al., 2016; Pino et al., 2020) have evaluated effectiveness of robot-assisted cognitive training on participant's cognitive capabilities. However, it is rarely addressed in the literature how acceptance and perception of the robot affects the effectiveness of cognitive training. Moreover, as shown in Table 2, some studies presented the results of robot-assisted training without comparing to the effectiveness of human-assisted training. For more rigorous evaluation of the effectiveness of a robot-assisted cognitive training approach, it is recommended to compare against human-assisted training and other existing approaches.

4.1.3. Uncontrollable Factors

There always exist uncontrollable factors during the study of robot-assisted cognitive training/rehabilitation. The problem is more noteworthy for multiple-session studies since researchers cannot control participants' daily and social activities outside of the laboratory setting. The topic of uncontrollable factors is relatively less studied. In a study on using a social robot to teach music to children with autism, Taheri et al. (2019) pointed out some improvements observed in music education and/or social skills are attributable to other interventions or education the participants may be receiving. When investigating the influence of robot-assisted training on cortical thickness in the brains of elderly participants, Kim et al. (2015) recognized uncontrollable factors due to participants' daily cognitive activity at home, such as using computers or reading books.

4.2. Challenges and Future Development

4.2.1. Ethical Challenges

During the development of robots for cognitive training/rehabilitation, there are some ethical issues with respect to human dignity, safety, legality, and social factors to be considered. For example, during robot-assisted cognitive training, the interaction between the user and the robot happens at both the cognitive (dominant) and physical level (Villaronga, 2016). There could be the issue of perceived safety, or cognitive harm. For example, the user may perceive the robot unsafe or scary (Salem et al., 2015; Coeckelbergh et al., 2016). In the study (Shim and Arkin, 2016) exploring the influence of robot deceptive behavior on human-robot interaction, a robot NAO deceptively showed positive feedback to participants' incorrect answers in a motor-cognition dual task. The self-report results revealed that the robot's deceptive feedback positively affected a human's frustration level and task engagement. Even though a robot's deceptive action may lead to positive outcome, Shim and Arkin emphasized that the ethical implications of the robot deception, including those regarding motives for deception, should always be discussed and validated prior to its application. Another arising ethical issue is how responsibility can be allocated, or distribution of responsibility (Loh, 2019; Müller, 2020). For example, if a robot acts during cognitive training, will the robot itself, designers or users be responsible, liable or accountable for the robot's actions? We should also pay close attention to ethical issues, such as the affective attachments, dependency on the robot, safety and privacy protection of users' information, and transparency in the use of algorithms in robotic systems (Kostavelis et al., 2015; Casey et al., 2016; Richardson et al., 2018; Fiske et al., 2019). Similarly, designers should accommodate the design of robot to these ethical considerations (Ozcana et al., 2016). To ensure the perceived safety, researchers need to always take end users' perception into account, which can be known through questionnaires and interviews with them (and their caregivers and therapists if needed), and their behavioral and neurophysiological activities. The tendency for humans to form attachments to anthropomorphized robots should be carefully considered during design (Riek and Howard, 2014; Riek, 2016). Moreover, for fear that the robot could replace human health care from both patients and the professional caregivers, it should be emphasized that the rehabilitation robots are developed with the aim of supplementing human caregivers, rather than replacing them (Doraiswamy et al., 2019).

4.2.2. User-Centered Design

The goal of robotic cognitive rehabilitation is to provide cost-effective cognitive training to vulnerable people with cognitive disabilities, which can supplement their caregivers and/or therapists (Doraiswamy et al., 2019). Therefore, we encourage the idea of user-driven, instead of technology-driven, robot design and development (Rehm et al., 2016). Emphasis should be given to the primary users (i.e., patients) of the robots and other key stakeholders (e.g., caregivers, therapists, and doctors) to design and shape this robot, including requirement analysis, robot development and evaluation with different stakeholders (Casey et al., 2016; Gerling et al., 2016; Leong and Johnston, 2016; Rehm et al., 2016; Salichs et al., 2016; Barco Martelo and Fosch Villaronga, 2017; Riek, 2017). It is also important to pay attention to potential technical difficulties for vulnerable populations, such as the elderly and children with ASD (Orejana et al., 2015) and the social and contextual environment that the robot will be applied to (Jones et al., 2015). More standardized, unbiased benchmarks and metrics need to be developed for different stakeholders to evaluate the performance of robots from their perspectives. While it is necessary to start pilot studies with healthy participants, it is crucial to relate the developed systems to patients with cognitive impairment at home settings.

Furthermore, robot development is a multidisciplinary study which requires knowledge from multiples fields, such as social cognitive science, engineering, psychology, and health care, such as ASD and dementia. Enhanced collaborations among these fields are needed to improve future technology in robotic rehabilitation.

4.2.3. Reliability, Trust, Transferability, and Cost-Effectiveness

The reliability of the robotic system ensures that the robot can consistently work in noisy, dynamic real-world environment. This makes a significant contribution to a user's confidence and increases positive perception of the robot (Wood et al., 2017). Mistakes made by robots during interaction can cause a loss of trust in human users. More work on human-robot interaction is needed to implement a reliable and robust robot to assist cognitive training. This may cover multimodal sensing technologies, artificial intelligence, and modeling. On the other hand, we need to take into consideration how to effectively restore trust to the robot in case that mistakes are made by the robot during interaction with the user. This may involve how and when the robot apologizes for its mistake for users. Robinette et al. (2015) found that timing is a key to repair robot trust, suggesting that the robot should wait and address the mistake next time a potential trust decision occurs rather than addressing the mistake immediately.

Currently, most studies focus on specific cognitive training tasks and environments, which means that the robot cannot assist in other cognitive tasks. Here, we encourage the implementation of a transferable mechanism in robots for cognitive training. In other words, we should enable more powerful learning algorithms in the robot so that the robot can learn and adapt to more new cognitive training (Andriella et al., 2019a). Researchers should also take cost-effectiveness into account during the design of robot (Wood et al., 2017). From the commercial perspective, cost-effectiveness is considered beyond the purchase, maintenance, and training costs for the system (Riek, 2017). Furthermore, from the perspective of time and effort of users, more work is needed to find out the optimal robot-assisted cognitive training strategy (e.g., the frequency and duration of each cognitive session). Therefore, we encourage future studies clearly state the used training strategies, making it easier for the community to compare different strategies.

4.2.4. Personalization

There is no “one-size-fits-all” in health care. To provide a successful cognitive training, the robot needs to be personalized and adaptive in three levels. Firstly, personalization requires the robot to provide appropriate cognitive training and feedback to meet the specific need of groups with different cognitive disabilities (e.g., people with ASD, people with ADRD). Secondly, the robots need to adapt to the diversity existing in the population as well as tailor to each individual user's severity of cognitive impairment, cultural and gender-dependent difference, personality and preference (Kostavelis et al., 2015; Javed et al., 2016; Riek, 2016; Darragh et al., 2017; Rudovic et al., 2017; Richardson et al., 2018; Sandygulova et al., 2019). For example, children with ASD, have a wide range of behavioral, social, and learning difficulties. And each individual may have a different preference to robot's gender and modalities of feedback (Sandygulova and O'Hare, 2015; Nault et al., 2020). As a result, we expect that a personalized robot would provide various cognitive training, e.g., a variety of games and adjustable voice, for diverse individual needs and requirements to keep the user engaged and focused over long term (Scassellati et al., 2018; Tleubayev et al., 2019). Furthermore, rehabilitation robots should adapt to individually time-changing characters, such as cognitive impairment, task engagement, and even personalities (Agrigoroaie and Tapus, 2016; Tsiakas et al., 2016, 2018). For example, the robot should adjust the cognitive training and feedback if the user feels bored, too difficult, or too easy (Lins et al., 2019). Machine learning methods should also take into consideration of personalization. Existing methods, such as interactive reinforcement learning (IRL) or incremental learning (Castellini, 2016) provide good examples, where one block module is used to specifically model each user's information, such as patient's name, hobbies and personalities related to cognitive training (Salichs et al., 2016). IRL is a variation of reinforcement learning that studies how a human can be included in the agent learning process. Human input play the role of feedback (i.e., reinforcement signal after the selected action) or guidance (i.e., actions to directly intervene/correct current strategy). IRL can also be utilized to enable adaptation and personalization during robot-assisted cognitive training. For example, Tsiakas et al. (2016) proposed an adaptive robot assisted cognitive therapy using IRL, where the primary user feedback input (e.g., engagement levels) were considered as a personalization factor and the guidance input from professional therapist were considered as a safety factor. Their simulation results showed that IRL improved the applied policy and led to a faster convergence to optimal policy. Castellini (2016) proposed an incremental learning model to enforce a true, endless adaptation of the robot to the subject and environment as well as improve the stability and reliability of robot's control. Incremental learning enables an adaptive robot system to update its own model whenever it is required, new information is available, or the prediction is deemed no longer reliable.

4.2.5. Human-Robot Collaboration

Future rehabilitation robots should not only be autonomous but also be collaborative (or co-operative) (Huijnen et al., 2016; Weiss et al., 2017). From the perspective of collaboration between the robot and the primary end users (i.e., people with cognitive disability), there is evidence indicating that a fully autonomous robotic system is not the best option for interaction with the vulnerable population (Peca et al., 2016). Instead, a semi-autonomous robot is a more suitable solution (Wood et al., 2017). With the highest-level goal of enhancing user's cognitive capabilities, the robot should “care” about the user's situation, take compensatory reaction as a teammate, engage the user and train/stimulate the user's cognitive capabilities as best as possible. The capability of collaboration may also help to avoid the user's feeling of redundancy and increase their feeling of self-autonomy and long-term engagement in cognitive training (Orejana et al., 2015). The robot should have a good perception of the user's changing situations and an intelligent strategy to engage the user. On the other hand, from the perspective of collaboration among robots, users, and their caregivers (and therapists), more future work is needed to solve the shared control issue. Researchers need to figure out strategies for robots to render the caregivers' and therapists' tasks easier as an assistive tool for cognitive training, instead of totally replacing them (Kostavelis et al., 2015; Coeckelbergh et al., 2016). A distribution between autonomy of robots and teleoperation by caregivers/therapists is needed to support the collaboration of robots for cognitive training.

4.2.6. Social Cognition

The knowledge gained in human-human interaction can be applied to foster human-robot interaction and to obtain critical insights for optimizing social encounters between humans and robots (Henschel et al., 2020). Marchesi et al. (2019) conducted a questionnaire study to investigate whether people adopt intentional stance (Dennett, 1989) toward a humanoid robot, iCub. Their results showed that it is possible to sometimes induce adoption of the intentional stance toward humanoid robots. Additionally, non-invasive neuroimaging techniques (e.g., fMRI) in neuroscience enable the possibility of probing social cognitive processing in human brain during interaction with robots. For example, Rauchbauer et al. (2019) observed that neural markers of mentalizing and social motivation were significantly more activated during human-human interaction than human-robot interaction. Klapper et al. (2014) showed that human brain activity within the theory-of-mind network (Saxe and Wexler, 2005; Price, 2012; Koster-Hale and Saxe, 2013) was reduced when interacting with agents in a form of low human animacy (i.e., physical appearance) compared to high human animacy. These issues become important when we adopt robots for cognitive training for people with cognitive dysfunction (Frith and Frith, 1999; Chevallier et al., 2012), as they underline the substantial contrast between human-human and human-agent interactions. Additionally, the advanced non-invasive, portable, and cost-effective neuroimaging techniques (e.g., EEG and fNIRS) hold the promise of evaluating human-robot interaction from controlled laboratory setting to real-world setting. Herein, we encourage to leverage human neuroscience to facilitate the development of robots for cognitive training, such as understanding the effects of robot-assisted cognitive training and learning the extent and contexts at which it can be beneficial from neurophysiological perspective.

4.2.7. Natural Human-Robot Interaction

Similar to human-human interaction during cognitive training by human therapists, robots need to be able to interact with users naturally in robotic rehabilitation. This includes having a good understanding of user's emotions (e.g., happiness, shame, engagement), intentions and personality (Pettinati and Arkin, 2015; Rahbar et al., 2015; Vaufreydaz et al., 2016; Rudovic et al., 2018), being able to provide an emotional response when being shared with personal information (de Graaf et al., 2015; Chumkamon et al., 2016), talking day-by-day more to the user on various topics like hobbies, and dealing with novel events (Dragone et al., 2015; Kostavelis et al., 2015; Adam et al., 2016; Ozcana et al., 2016). These natural user-robot interactions require powerful perception, reasoning, acting and learning modules in robots, or in other words, cognitive and social-emotional capabilities. However, from the perception perspective, understanding users' intentions and emotions is still a great challenge for robots (Matarić, 2017). Robots need to interpret multimodal signals (e.g., facial expression, gestures, voice, and speech) simultaneously to understand users' covert intentions and emotions. Similarly, more work is needed for the multimodal feedback in the future. To maximize the benefits of physically present robots and facilitate both the short- and long-term human-robot interaction for cognitive training, we need to develop more embodied communication in robots, not limited in verbal communication (Paradeda et al., 2016; Salichs et al., 2016; Matarić, 2017). Haptic sensing and feedback should be strongly considered in future research as part of multimodal perception and feedback (Arnold and Scheutz, 2017; Cangelosi and Invitto, 2017). More specifically, we need to implement the strategy to enable the robot to associate cognitive assistance and exercise with appropriate multimodal feedback, e.g., spoken words, facial expressions, eye gaze, and other body movements (Paletta et al., 2018). The embodied communication during human-robot interaction is a challenging research area (Nunez et al., 2015). It is still unclear how and how much the embodied communication from the robot can influence user's perception of the robot (Dubois et al., 2016). Moreover, previous studies indicated that users' experience of the robot could also be influenced by the unexpected behaviors (Lemaignan et al., 2015), synchrony and reciprocity (Lorenz et al., 2016), and even cognitive biases (Biswas and Murray, 2016, 2017) from the robot. A caveat is that there still exist many unknowns for natural human-robot interaction.

In summary, to achieve natural human-robot interaction during cognitive training requires not only multimodal sensing technology and artificial intelligence (e.g., deep learning) (Jing et al., 2015; Lopez-Samaniego and Garcia-Zapirain, 2016; Pierson and Gashler, 2017) but also the development of related fields (Wan et al., 2020), such as cognitive computing (Chen et al., 2018), social-cognitive mechanisms (Wiltshire et al., 2017), and modeling of cognitive architectures (Kotseruba et al., 2016; Woo et al., 2017).

5. Conclusion

Robot-assisted cognitive training is becoming an affordable promise for people with cognitive disabilities. In this review paper, we present a systematic review on the current application, enabling technologies, and main commercial robots in the field of robot-assisted cognitive training. Many studies have been successfully conducted to evaluate the feasibility, safety, usability, and effectiveness of robotic rehabilitation. Existing studies often include a small sample size. Also, the questionnaires need to be standardized both to evaluate the overall experience with the robot and the impact of the robot on the specific cognitive ability that it aims to assess. There are still multifaceted challenges in the application of robots in cognitive training. Firstly, ethical issues, such as human safety and violation of social norms, can arise during robot-assisted cognitive training. Secondly, with respect to the design of a robot-assisted cognitive training system, the developers should have a close collaboration with the end-users and stakeholders from the initial design, implementation, evaluation and improvement. Thirdly, the trust, reliability and the cost-effectiveness should be taken into account. Moreover, the rehabilitation robot should be capable to adapt and personalize to the specific individual need, and also learn to collaborate with users in the future. The recent advancement of social cognition may facilitate the human-robot interaction during cognitive training. Lastly, the rehabilitation robot should be able to interact with users in a natural way, similar to the human-human interaction during cognitive training. Noticeably, these challenges are mutually influencing one another. Cross-disciplinary collaboration is necessary to solve these challenges in future.

Data Availability Statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author/s.

Author Contributions

FY, RL, and XZ determined the review scope and review strategies. FY, EK, and ZL conducted the searching and screening of the literature, and reviewing of the selected articles. FY, EK, RL, and XZ wrote the manuscript. All authors contributed to the article and approved the submitted version.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

Ab Aziz, A., Ahmad, F., Yusof, N., Ahmad, F. K., and Yusof, S. A. M. (2015). “Designing a robot-assisted therapy for individuals with anxiety traits and states,” in 2015 International Symposium on Agents, Multi-Agent Systems and Robotics (ISAMSR) (Putrajaya: IEEE), 98–103. doi: 10.1109/ISAMSR.2015.7379778

CrossRef Full Text | Google Scholar

Abdollahi, H., Mollahosseini, A., Lane, J. T., and Mahoor, M. H. (2017). “A pilot study on using an intelligent life-like robot as a companion for elderly individuals with dementia and depression,” in 2017 IEEE-RAS 17th International Conference on Humanoid Robotics (Humanoids) (Birmingham: IEEE), 541–546. doi: 10.1109/HUMANOIDS.2017.8246925

CrossRef Full Text | Google Scholar

Adam, C., Johal, W., Pellier, D., Fiorino, H., and Pesty, S. (2016). “Social human-robot interaction: a new cognitive and affective interaction-oriented architecture,” in International Conference on Social Robotics (Springer), 253–263. doi: 10.1007/978-3-319-44884-8

CrossRef Full Text | Google Scholar

Agrigoroaie, R., Ferland, F., and Tapus, A. (2016). The enrichme project: Lessons learnt from a first interaction with the elderly,” in International Conference on Social Robotics (Kansas City, MO: Springer), 735–745. doi: 10.1007/978-3-319-47437-3_72

CrossRef Full Text | Google Scholar

Agrigoroaie, R. M., and Tapus, A. (2016). “Developing a healthcare robot with personalized behaviors and social skills for the elderly,” in 2016 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI) (Christchurch: IEEE), 589–590. doi: 10.1109/HRI.2016.7451870

CrossRef Full Text | Google Scholar

Ahn, H. S., Lee, M. H., Broadbent, E., and MacDonald, B. A. (2017). “Is entertainment services of a healthcare service robot for older people useful to young people?” in 2017 First IEEE International Conference on Robotic Computing (IRC) (Taichung: IEEE), 330–335. doi: 10.1109/IRC.2017.70

CrossRef Full Text | Google Scholar

Alemi, M., Ghanbarzadeh, A., Meghdari, A., and Moghadam, L. J. (2016). Clinical application of a humanoid robot in pediatric cancer interventions. Int. J. Soc. Robot. 8, 743–759. doi: 10.1007/s12369-015-0294-y

CrossRef Full Text | Google Scholar

Ali, S., Mehmood, F., Dancey, D., Ayaz, Y., Khan, M. J., Naseer, N., et al. (2019). An adaptive multi-robot therapy for improving joint attention and imitation of asd children. IEEE Access 7, 81808–81825. doi: 10.1109/ACCESS.2019.2923678

CrossRef Full Text | Google Scholar

Alimardani, M., and Hiraki, K. (2017). Development of a real-time brain-computer interface for interactive robot therapy: an exploration of eeg and emg features during hypnosis. Int. J. Comput. Inform. Eng. 11, 187–195. doi: 10.5281/zenodo.1340062

CrossRef Full Text | Google Scholar

Alnajjar, F., Cappuccio, M., Renawi, A., Mubin, O., and Loo, C. K. (2020). Personalized robot interventions for autistic children: an automated methodology for attention assessment. Int. J. Soc. Robot. 13, 67–82. doi: 10.1007/s12369-020-00639-8

CrossRef Full Text | Google Scholar

Aminuddin, R., Sharkey, A., and Levita, L. (2016). “Interaction with the paro robot may reduce psychophysiological stress responses,” in 2016 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI) (Christchurch: IEEE), 593–594. doi: 10.1109/HRI.2016.7451872

CrossRef Full Text | Google Scholar

Andriella, A., Suárez-Hernández, A., Segovia-Aguas, J., Torras, C., and Alenyà, G. (2019a). “Natural teaching of robot-assisted rearranging exercises for cognitive training,” in International Conference on Social Robotics (Madrid: Springer), 611–621. doi: 10.1007/978-3-030-35888-4_57

CrossRef Full Text | Google Scholar

Andriella, A., Torras, C., and Alenyà, G. (2019b). “Learning robot policies using a high-level abstraction persona-behaviour simulator,” in 2019 28th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN) (New Delhi: IEEE), 1–8. doi: 10.1109/RO-MAN46459.2019.8956357

CrossRef Full Text | Google Scholar

Andriella, A., Torras, C., and Alenya, G. (2020). Cognitive system framework for brain-training exercise based on human-robot interaction. Cogn. Comput. 12, 793–810. doi: 10.1007/s12559-019-09696-2

CrossRef Full Text | Google Scholar

Andruseac, G. G., Adochiei, R. I., Păsărică, A., Adochiei, F. C., Corciovă, C., and Costin, H. (2015). “Training program for dyslexic children using educational robotics,” in 2015 E-Health and Bioengineering Conference (EHB) (Iasi: IEEE), 1–4. doi: 10.1109/EHB.2015.7391547

CrossRef Full Text | Google Scholar

Ansado, J., Chasen, C., Bouchard, S., Iakimova, G., and Northoff, G. (2020). How brain imaging provides predictive biomarkers for therapeutic success in the context of virtual reality cognitive training. Neurosci. Biobehav. Rev. 120, 583–594. doi: 10.1016/j.neubiorev.2020.05.018

PubMed Abstract | CrossRef Full Text | Google Scholar

Arnold, T., and Scheutz, M. (2017). The tactile ethics of soft robotics: designing wisely for human–robot interaction. Soft Robot. 4, 81–87. doi: 10.1089/soro.2017.0032

PubMed Abstract | CrossRef Full Text | Google Scholar

Barco Martelo, A., and Fosch Villaronga, E. (2017). “Child-robot interaction studies: from lessons learned to guidelines,” in 3rd Workshop on Child-Robot Interaction, CRI 2017 (Vienna).

Google Scholar

Beckerle, P., Salvietti, G., Unal, R., Prattichizzo, D., Rossi, S., Castellini, C., et al. (2017). A human-robot interaction perspective on assistive and rehabilitation robotics. Front. Neurorobot. 11:24. doi: 10.3389/fnbot.2017.00024

PubMed Abstract | CrossRef Full Text | Google Scholar

Begum, M., Serna, R. W., Kontak, D., Allspaw, J., Kuczynski, J., Yanco, H. A., et al. (2015). “Measuring the efficacy of robots in autism therapy: how informative are standard HRI metrics?” in Proceedings of the Tenth Annual ACM/IEEE International Conference on Human-Robot Interaction (Portland, OR), 335–342. doi: 10.1145/2696454.2696480

CrossRef Full Text | Google Scholar

Bertani, R., Melegari, C., Maria, C., Bramanti, A., Bramanti, P., and Calabrò, R. S. (2017). Effects of robot-assisted upper limb rehabilitation in stroke patients: a systematic review with meta-analysis. Neurol. Sci. 38, 1561–1569. doi: 10.1007/s10072-017-2995-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Bharatharaj, J., Huang, L., Mohan, R. E., Al-Jumaily, A., and Krägeloh, C. (2017). Robot-assisted therapy for learning and social interaction of children with autism spectrum disorder. Robotics 6:4. doi: 10.3390/robotics6010004

CrossRef Full Text | Google Scholar

Biswas, M., and Murray, J. (2016). “The effects of cognitive biases in long-term human-robot interactions: case studies using three cognitive biases on marc the humanoid robot,” in International Conference on Social Robotics (Kansas City, MO: Springer), 148–158. doi: 10.1007/978-3-319-47437-3_15

CrossRef Full Text | Google Scholar

Biswas, M., and Murray, J. (2017). The effects of cognitive biases and imperfectness in long-term robot-human interactions: case studies using five cognitive biases on three robots. Cogn. Syst. Res. 43, 266–290. doi: 10.1016/j.cogsys.2016.07.007

CrossRef Full Text | Google Scholar

Botvinick, M., Ritter, S., Wang, J. X., Kurth-Nelson, Z., Blundell, C., and Hassabis, D. (2019). Reinforcement learning, fast and slow. Trends Cogn. Sci. 23, 408–422. doi: 10.1016/j.tics.2019.02.006

CrossRef Full Text | Google Scholar

Bozgeyikli, L., Bozgeyikli, E., Clevenger, M., Raij, A., Alqasemi, R., Sundarrao, S., et al. (2015). “Vr4vr: vocational rehabilitation of individuals with disabilities in immersive virtual reality environments,” in Proceedings of the 8th ACM International Conference on PErvasive Technologies Related to Assistive Environments (Corfu), 1–4. doi: 10.1145/2769493.2769592

CrossRef Full Text | Google Scholar

Calderita, L. V., Vega, A., Barroso-Ramírez, S., Bustos, P., and Núñez, P. (2020). Designing a cyber-physical system for ambient assisted living: a use-case analysis for social robot navigation in caregiving centers. Sensors 20:4005. doi: 10.3390/s20144005

PubMed Abstract | CrossRef Full Text | Google Scholar

Cangelosi, A., and Invitto, S. (2017). Human-robot interaction and neuroprosthetics: a review of new technologies. IEEE Consum. Electron. Mag. 6, 24–33. doi: 10.1109/MCE.2016.2614423

CrossRef Full Text | Google Scholar

Cao, H. L., Pop, C., Simut, R., Furnemónt, R., De Beir, A., Van de Perre, G., et al. (2015). “Probolino: a portable low-cost social device for home-based autism therapy,” in International Conference on Social Robotics (Paris: Springer), 93–102. doi: 10.1007/978-3-319-25554-5_10

CrossRef Full Text | Google Scholar

Carros, F., Meurer, J., Löffler, D., Unbehaun, D., Matthies, S., Koch, I., et al. (2020). “Exploring human-robot interaction with the elderly: results from a ten-week case study in a care home,” in Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (Honolulu, HI), 1–12. doi: 10.1145/3313831.3376402

CrossRef Full Text | Google Scholar

Casey, D., Felzmann, H., Pegman, G., Kouroupetroglou, C., Murphy, K., Koumpis, A., et al. (2016). “What people with dementia want: designing mario an acceptable robot companion,” in International Conference on Computers Helping People With Special Needs, (Linz: Springer), 318–325. doi: 10.1007/978-3-319-41264-1_44

CrossRef Full Text | Google Scholar

Castellini, C. (2016). “Incremental learning of muscle synergies: from calibration to interaction,” in Human and Robot Hands (Cham: Springer), 171–193. doi: 10.1007/978-3-319-26706-7_11

CrossRef Full Text | Google Scholar

Chen, H., Zhu, H., Teng, Z., and Zhao, P. (2020). Design of a robotic rehabilitation system for mild cognitive impairment based on computer vision. J. Eng. Sci. Med. Diagn. Ther. 3:021108. doi: 10.1115/1.4046396

CrossRef Full Text | Google Scholar

Chen, M., Herrera, F., and Hwang, K. (2018). Cognitive computing: architecture, technologies and intelligent applications. IEEE Access 6, 19774–19783. doi: 10.1109/ACCESS.2018.2791469

CrossRef Full Text | Google Scholar

Chevallier, C., Kohls, G., Troiani, V., Brodkin, E. S., and Schultz, R. T. (2012). The social motivation theory of autism. Trends Cogn. Sci. 16, 231–239. doi: 10.1016/j.tics.2012.02.007

CrossRef Full Text | Google Scholar

Chu, M. T., Khosla, R., Khaksar, S. M. S., and Nguyen, K. (2017). Service innovation through social robot engagement to improve dementia care quality. Assist. Technol. 29, 8–18. doi: 10.1080/10400435.2016.1171807

PubMed Abstract | CrossRef Full Text | Google Scholar

Chumkamon, S., Hayashi, E., and Koike, M. (2016). Intelligent emotion and behavior based on topological consciousness and adaptive resonance theory in a companion robot. Biol. Inspir. Cogn. Architect. 18, 51–67. doi: 10.1016/j.bica.2016.09.004

CrossRef Full Text | Google Scholar

Clabaugh, C. E., Mahajan, K., Jain, S., Pakkar, R., Becerra, D., Shi, Z., et al. (2019). Long-term personalization of an in-home socially assistive robot for children with autism spectrum disorders. Front. Robot. AI 6:110. doi: 10.3389/frobt.2019.00110

PubMed Abstract | CrossRef Full Text | Google Scholar

Coeckelbergh, M., Pop, C., Simut, R., Peca, A., Pintea, S., David, D., et al. (2016). A survey of expectations about the role of robots in robot-assisted therapy for children with ASD: ethical acceptability, trust, sociability, appearance, and attachment. Sci. Eng. Ethics 22, 47–65. doi: 10.1007/s11948-015-9649-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Conti, D., Di Nuovo, S., Buono, S., Trubia, G., and Di Nuovo, A. (2015). “Use of robotics to stimulate imitation in children with autism spectrum disorder: A pilot study in a clinical setting,” in 2015 24th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN) (Kobe: IEEE), 1–6. doi: 10.1109/ROMAN.2015.7333589

CrossRef Full Text | Google Scholar

Costescu, C. A., Vanderborght, B., and David, D. O. (2015). Reversal learning task in children with autism spectrum disorder: a robot-based approach. J. Autism Dev. Disord. 45, 3715–3725. doi: 10.1007/s10803-014-2319-z

PubMed Abstract | CrossRef Full Text | Google Scholar

D'Amico, A., and Guastella, D. (2019). The robotic construction kit as a tool for cognitive stimulation in children and adolescents: the re4bes protocol. Robotics 8:8. doi: 10.3390/robotics8010008

CrossRef Full Text | Google Scholar

Darragh, M., Ahn, H. S., MacDonald, B., Liang, A., Peri, K., Kerse, N., et al. (2017). Homecare robots to improve health and well-being in mild cognitive impairment and early stage dementia: results from a scoping study. J. Ame. Med. Direc. Assoc. 18:1099-e1. doi: 10.1016/j.jamda.2017.08.019

PubMed Abstract | CrossRef Full Text | Google Scholar

David, D. O., Costescu, C. A., Matu, S., Szentagotai, A., and Dobrean, A. (2018). Developing joint attention for children with autism in robot-enhanced therapy. Int. J. Soc. Robot. 10, 595–605. doi: 10.1007/s12369-017-0457-0

CrossRef Full Text | Google Scholar

David, D. O., Costescu, C. A., Matu, S., Szentagotai, A., and Dobrean, A. (2020). Effects of a robot-enhanced intervention for children with ASD on teaching turn-taking skills. J. Educ. Comput. Res. 58, 29–62. doi: 10.1177/0735633119830344

CrossRef Full Text | Google Scholar

de Graaf, M. M., Allouch, S. B., and Van Dijk, J. (2015). “What makes robots social? A user's perspective on characteristics for social human-robot interaction,” in International Conference on Social Robotics (Paris: Springer), 184–193. doi: 10.1007/978-3-319-25554-5_19

CrossRef Full Text | Google Scholar

Demetriadis, S., Tsiatsos, T., Sapounidis, T., Tsolaki, M., and Gerontidis, A. (2016). “Exploring the potential of programming tasks to benefit patients with mild cognitive impairment,” in Proceedings of the 9th ACM International Conference on PErvasive Technologies Related to Assistive Environments (Corfu Island Greece), 1–4. doi: 10.1145/2910674.2935850

CrossRef Full Text | Google Scholar

Dennett, D. C. (1989). The Intentional Stance. Cambridge, MA: MIT Press.

Google Scholar

Dickerson, B. C. (2015). Dysfunction of social cognition and behavior. Continuum 21:660. doi: 10.1212/01.CON.0000466659.05156.1d

CrossRef Full Text | Google Scholar

Doraiswamy, P. M., London, E., Varnum, P., Harvey, B., Saxena, S., Tottman, S., et al. (2019). Empowering 8 Billion Minds: Enabling Better Mental Health for All via the Ethical Adoption of Technologies. Washington, DC: NAM Perspectives.

Google Scholar

Dragone, M., Amato, G., Bacciu, D., Chessa, S., Coleman, S., Di Rocco, M., et al. (2015). A cognitive robotic ecology approach to self-configuring and evolving aal systems. Eng. Appl. Artif. Intell. 45, 269–280. doi: 10.1016/j.engappai.2015.07.004

CrossRef Full Text | Google Scholar

Dubois, M., Claret, J. A., Basañez, L., and Venture, G. (2016). “Influence of emotional motions in human-robot interactions,” in International Symposium on Experimental Robotics (Tokyo: Springer), 799–808. doi: 10.1007/978-3-319-50115-4_69

CrossRef Full Text | Google Scholar

Esteban, P. G., Baxter, P., Belpaeme, T., Billing, E., Cai, H., Cao, H. L., et al. (2017). How to build a supervised autonomous system for robot-enhanced therapy for children with autism spectrum disorder. Paladyn J. Behav. Robot. 8, 18–38. doi: 10.1515/pjbr-2017-0002

CrossRef Full Text

Fiske, A., Henningsen, P., and Buyx, A. (2019). Your robot therapist will see you now: ethical implications of embodied artificial intelligence in psychiatry, psychology, and psychotherapy. J. Med. Internet Res. 21:e13216. doi: 10.2196/13216

PubMed Abstract | CrossRef Full Text | Google Scholar

Frith, C. D., and Frith, U. (1999). Interacting minds-a biological basis. Science 286, 1692–1695. doi: 10.1126/science.286.5445.1692

CrossRef Full Text | Google Scholar

Garcia-Sanjuan, F., Jaen, J., and Nacher, V. (2017). Tangibot: a tangible-mediated robot to support cognitive games for ageing people—a usability study. Perv. Mob. Comput. 34, 91–105. doi: 10.1016/j.pmcj.2016.08.007

CrossRef Full Text | Google Scholar

Gerling, K., Hebesberger, D., Dondrup, C., Körtner, T., and Hanheide, M. (2016). Robot deployment in long-term care. Z. Gerontol. Geriatr. 49, 288–297. doi: 10.1007/s00391-016-1065-6

CrossRef Full Text | Google Scholar

Hajat, C., and Stein, E. (2018). The global burden of multiple chronic conditions: a narrative review. Prev. Med. Rep. 12, 284–293. doi: 10.1016/j.pmedr.2018.10.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Haring, K. S., Silvera-Tawil, D., Watanabe, K., and Velonaki, M. (2016). “The influence of robot appearance and interactive ability in HRI: a cross-cultural study,” in International Conference on Social Robotics (Kansas City, MO: Springer), 392–401. doi: 10.1007/978-3-319-47437-3_38

CrossRef Full Text | Google Scholar

Hartson, R., and Pyla, P. S. (2018). The UX Book: Agile UX Design for a Quality User Experience. Cambridge, MA: Morgan Kaufmann.

Google Scholar

Havins, W. N., Massman, P. J., and Doody, R. (2012). Factor structure of the geriatric depression scale and relationships with cognition and function in Alzheimer's disease. Dement. Geriatr. Cogn. Disord. 34, 360–372. doi: 10.1159/000345787

PubMed Abstract | CrossRef Full Text | Google Scholar

Heins, S., Dehem, S., Montedoro, V., Dehez, B., Edwards, M., Stoquart, G., et al. (2017). “Robotic-assisted serious game for motor and cognitive post-stroke rehabilitation,” in 2017 IEEE 5th International Conference on Serious Games and Applications for Health (SeGAH) (Perth, WA: IEEE), 1–8. doi: 10.1109/SeGAH.2017.7939262

CrossRef Full Text | Google Scholar

Henschel, A., Hortensius, R., and Cross, E. S. (2020). Social cognition in the age of human-robot interaction. Trends Neurosci. 43, 373–384. doi: 10.1016/j.tins.2020.03.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Huijnen, C. A., Lexis, M. A., and de Witte, L. P. (2016). Matching robot kaspar to autism spectrum disorder (ASD) therapy and educational goals. Int. J. Soc. Robot. 8, 445–455. doi: 10.1007/s12369-016-0369-4

CrossRef Full Text | Google Scholar

Huskens, B., Palmen, A., Van der Werff, M., Lourens, T., and Barakova, E. (2015). Improving collaborative play between children with autism spectrum disorders and their siblings: the effectiveness of a robot-mediated intervention based on Lego® therapy. J. Autism Dev. Disord. 45, 3746–3755. doi: 10.1007/s10803-014-2326-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Ioannou, A., Kartapanis, I., and Zaphiris, P. (2015). “Social robots as co-therapists in autism therapy sessions: a single-case study,” in International Conference on Social Robotics (Paris: Springer), 255–263. doi: 10.1007/978-3-319-25554-5_26

CrossRef Full Text | Google Scholar

Javed, H., Cabibihan, J. J., Aldosari, M., and Al-Attiyah, A. (2016). “Culture as a driver for the design of social robots for autism spectrum disorder interventions in the Middle East,” in International Conference on Social Robotics (Kansas City, MO: Springer), 591–599. doi: 10.1007/978-3-319-47437-3_58

CrossRef Full Text | Google Scholar

Javed, H., Jeon, M., and Park, C. H. (2018). “Adaptive framework for emotional engagement in child-robot interactions for autism interventions,” in 2018 15th International Conference on Ubiquitous Robots (UR) (Honolulu, HI: IEEE), 396–400. doi: 10.1109/URAI.2018.8441775

CrossRef Full Text | Google Scholar

Jing, H., Lun, X., Dan, L., Zhijie, H., and Zhiliang, W. (2015). Cognitive emotion model for eldercare robot in smart home. China Commun. 12, 32–41. doi: 10.1109/CC.2015.7114067

CrossRef Full Text | Google Scholar

Johal, W., Calvary, G., and Pesty, S. (2015). “Non-verbal signals in HRI: interference in human perception,” in International Conference on Social Robotics (Paris: Springer), 275–284. doi: 10.1007/978-3-319-25554-5_28

CrossRef Full Text | Google Scholar

Jones, A., Küster, D., Basedow, C. A., Alves-Oliveira, P., Serholt, S., Hastie, H., et al. (2015). “Empathic robotic tutors for personalised learning: a multidisciplinary approach,” in International Conference on Social Robotics (Paris: Springer), 285–295. doi: 10.1007/978-3-319-25554-5_29

CrossRef Full Text | Google Scholar

Kajopoulos, J., Wong, A. H. Y., Yuen, A. W. C., Dung, T. A., Kee, T. Y., and Wykowska, A. (2015). “Robot-assisted training of joint attention skills in children diagnosed with autism,” in International Conference on Social Robotics (Paris: Springer), 296–305. doi: 10.1007/978-3-319-25554-5_30

CrossRef Full Text | Google Scholar

Khosla, R., Nguyen, K., and Chu, M. T. (2017). Human robot engagement and acceptability in residential aged care. Int. J. Hum. Comput. Interact. 33, 510–522. doi: 10.1080/10447318.2016.1275435

CrossRef Full Text | Google Scholar

Kim, G., Lim, S., Kim, H., Lee, B., Seo, S., Cho, K., et al. (2017). Is robot-assisted therapy effective in upper extremity recovery in early stage stroke?—A systematic literature review. J. Phys. Ther. Sci. 29, 1108–1112. doi: 10.1589/jpts.29.1108

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, G. H., Jeon, S., Im, K., Kwon, H., Lee, B. H., Kim, G. Y., et al. (2015). Structural brain changes after traditional and robot-assisted multi-domain cognitive training in community-dwelling healthy elderly. PLoS ONE 10:e0123251. doi: 10.1371/journal.pone.0123251

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, G. H., Kim, B., and Jeong, J. H. (2019). The effects of home-based cognitive intervention with robots on cognitive performance in patients with mild cognitive impairment. Alzheimers Dement. 15:P154. doi: 10.1016/j.jalz.2019.06.4314

CrossRef Full Text | Google Scholar

Klapper, A., Ramsey, R., Wigboldus, D., and Cross, E. S. (2014). The control of automatic imitation based on bottom-up and top-down cues to animacy: insights from brain and behavior. J. Cogn. Neurosci. 26, 2503–2513. doi: 10.1162/jocn_a_00651

PubMed Abstract | CrossRef Full Text | Google Scholar

Kohori, T., Hirayama, S., Hara, T., Muramatsu, M., Naganuma, H., Yamano, M., et al. (2017). “Development and evaluation of an interactive therapy robot,” in International Conference on Advances in Computer Entertainment (London: Springer), 66–83. doi: 10.1007/978-3-319-76270-8_6

CrossRef Full Text | Google Scholar

Korchut, A., Szklener, S., Abdelnour, C., Tantinya, N., Hernández-Farigola, J., Ribes, J. C., et al. (2017). challenges for service robots—requirements of elderly adults with cognitive impairments. Front. Neurol. 8:228. doi: 10.3389/fneur.2017.00228

PubMed Abstract | CrossRef Full Text | Google Scholar

Kostavelis, I., Giakoumis, D., Malasiotis, S., and Tzovaras, D. (2015). “Ramcip: towards a robotic assistant to support elderly with mild cognitive impairments at home,” in International Symposium on Pervasive Computing Paradigms for Mental Health (Milan: Springer), 186–195. doi: 10.1007/978-3-319-32270-4_19

CrossRef Full Text | Google Scholar

Koster-Hale, J., and Saxe, R. (2013). Theory of mind: a neural prediction problem. Neuron 79, 836–848. doi: 10.1016/j.neuron.2013.08.020

PubMed Abstract | CrossRef Full Text | Google Scholar

Kotseruba, I., Gonzalez, O. J. A., and Tsotsos, J. K. (2016). A review of 40 years of cognitive architecture research: focus on perception, attention, learning and applications. arXiv 1610.08602.

Google Scholar

Law, M., Ahn, H. S., MacDonald, B., Vasjakovic, D. S., Lim, J., Lee, M. H., et al. (2019a). “User testing of cognitive training games for people with mild cognitive impairment: design implications,” in International Conference on Social Robotics (Madrid: Springer), 464–473. doi: 10.1007/978-3-030-35888-4_43

CrossRef Full Text | Google Scholar

Law, M., Sutherland, C., Ahn, H. S., MacDonald, B. A., Peri, K., Johanson, D. L., et al. (2019b). Developing assistive robots for people with mild cognitive impairment and mild dementia: a qualitative study with older adults and experts in aged care. BMJ Open 9:e031937. doi: 10.1136/bmjopen-2019-031937

PubMed Abstract | CrossRef Full Text | Google Scholar

Lemaignan, S., Fink, J., Mondada, F., and Dillenbourg, P. (2015). “You're doing it wrong! studying unexpected behaviors in child-robot interaction,” in International Conference on Social Robotics (Paris: Springer), 390–400. doi: 10.1007/978-3-319-25554-5_39

CrossRef Full Text | Google Scholar

Leong, T. W., and Johnston, B. (2016). Co-design and robots: a case study of a robot dog for aging people,” in International Conference on Social Robotics (Kansas City, MO: Springer), 702–711. doi: 10.1007/978-3-319-47437-3_69

CrossRef Full Text | Google Scholar

Li, J. (2015). The benefit of being physically present: a survey of experimental works comparing copresent robots, telepresent robots and virtual agents. Int. J. Hum. Comput. Stud. 77, 23–37. doi: 10.1016/j.ijhcs.2015.01.001

CrossRef Full Text | Google Scholar

Lins, A. A., de Oliveira, J. M., Rodrigues, J. J., and de Albuquerque, V. H. C. (2019). Robot-assisted therapy for rehabilitation of children with cerebral palsy-a complementary and alternative approach. Comput. Hum. Behav. 100, 152–167. doi: 10.1016/j.chb.2018.05.012

CrossRef Full Text | Google Scholar

Loh, J. (2019). Responsibility and robot ethics: a critical overview. Philosophies 4:58. doi: 10.3390/philosophies4040058

CrossRef Full Text | Google Scholar

Lopez-Samaniego, L., and Garcia-Zapirain, B. (2016). A robot-based tool for physical and cognitive rehabilitation of elderly people using biofeedback. Int. J. Environ. Res. Public Health 13:1176. doi: 10.3390/ijerph13121176

PubMed Abstract | CrossRef Full Text | Google Scholar

LoPresti, E. F., Bodine, C., and Lewis, C. (2008). Assistive technology for cognition [understanding the needs of persons with disabilities]. IEEE Eng. Med. Biol. Mag. 27, 29–39. doi: 10.1109/EMB.2007.907396

CrossRef Full Text | Google Scholar

Lorenz, T., Weiss, A., and Hirche, S. (2016). Synchrony and reciprocity: key mechanisms for social companion robots in therapy and care. Int. J. Soc. Robot. 8, 125–143. doi: 10.1007/s12369-015-0325-8

CrossRef Full Text | Google Scholar

Manca, M., Paternò, F., Santoro, C., Zedda, E., Braschi, C., Franco, R., et al. (2020). The impact of serious games with humanoid robots on mild cognitive impairment older adults. Int. J. Hum. Comput. Stud. 145:102509. doi: 10.1016/j.ijhcs.2020.102509

CrossRef Full Text | Google Scholar

Marchesi, S., Ghiglino, D., Ciardo, F., Perez-Osorio, J., Baykara, E., and Wykowska, A. (2019). Do we adopt the intentional stance toward humanoid robots? Front. Psychol. 10:450. doi: 10.3389/fpsyg.2019.00450

PubMed Abstract | CrossRef Full Text | Google Scholar

Marino, F., Chilà, P., Sfrazzetto, S. T., Carrozza, C., Crimi, I., Failla, C., et al. (2019). Outcomes of a robot-assisted social-emotional understanding intervention for young children with autism spectrum disorders. J. Autism Dev. Disord. 50, 1973–1987. doi: 10.1007/s10803-019-03953-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Matarić, M. J. (2017). Socially assistive robotics: human augmentation versus automation. Sci. Robot. 2:eaam5410. doi: 10.1126/scirobotics.aam5410

PubMed Abstract | CrossRef Full Text | Google Scholar

Melo, F. S., Sardinha, A., Belo, D., Couto, M., Faria, M., Farias, A., et al. (2019). Project inside: towards autonomous semi-unstructured human-robot social interaction in autism therapy. Artif. Intell. Med. 96, 198–216. doi: 10.1016/j.artmed.2018.12.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Mewborn, C. M., Lindbergh, C. A., and Miller, L. S. (2017). Cognitive interventions for cognitively healthy, mildly impaired, and mixed samples of older adults: a systematic review and meta-analysis of randomized-controlled trials. Neuropsychol. Rev. 27, 403–439. doi: 10.1007/s11065-017-9350-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Miskam, M. A., Shamsuddin, S., Yussof, H., Ariffin, I. M., and Omar, A. R. (2015). “A questionnaire-based survey: therapist's response on emotions gestures using humanoid robot for autism,” in 2015 International Symposium on Micro-NanoMechatronics and Human Science (MHS) (Nagoya: IEEE), 1–7. doi: 10.1109/MHS.2015.7438298

CrossRef Full Text | Google Scholar

Mois, G., Collete, B. A., Renzi-Hammond, L. M., Boccanfuso, L., Ramachandran, A., Gibson, P., et al. (2020). “Understanding robots' potential to facilitate piano cognitive training in older adults with mild cognitive impairment,” in Companion of the 2020 ACM/IEEE International Conference on Human-Robot Interaction (Cambridge), 363–365. doi: 10.1145/3371382.3378299

CrossRef Full Text | Google Scholar

Mori, M., MacDorman, K. F., and Kageki, N. (2012). The uncanny valley [from the field]. IEEE Robot. Autom. Mag. 19, 98–100. doi: 10.1109/MRA.2012.2192811

CrossRef Full Text | Google Scholar

Morone, G., Paolucci, S., Cherubini, A., De Angelis, D., Venturiero, V., Coiro, P., et al. (2017). Robot-assisted gait training for stroke patients: current state of the art and perspectives of robotics. Neuropsychiatr. Dis. Treat. 13:1303. doi: 10.2147/NDT.S114102

PubMed Abstract | CrossRef Full Text | Google Scholar

Müller, V. C. (2020). “Ethics of artificial intelligence and robotics,” in The Stanford Encyclopedia of Philosophy, Winter 2020 Edn., ed. E. N. Zalta (Stanford, CA: Metaphysics Research Lab, Stanford University).

PubMed Abstract | Google Scholar

Nault, E., Baillie, L., and Broz, F. (2020). “Auditory and haptic feedback in a socially assistive robot memory game,” in Companion of the 2020 ACM/IEEE International Conference on Human-Robot Interaction (Cambridge), 369–371. doi: 10.1145/3371382.3378375

CrossRef Full Text | Google Scholar

Nunez, E., Matsuda, S., Hirokawa, M., and Suzuki, K. (2015). “Humanoid robot assisted training for facial expressions recognition based on affective feedback,” in International Conference on Social Robotics (Paris: Springer), 492–501. doi: 10.1007/978-3-319-25554-5_49

CrossRef Full Text | Google Scholar

Orejana, J. R., MacDonald, B. A., Ahn, H. S., Peri, K., and Broadbent, E. (2015). “Healthcare robots in homes of rural older adults,” in International Conference on Social Robotics (Paris: Springer), 512–521. doi: 10.1007/978-3-319-25554-5_51

CrossRef Full Text | Google Scholar

Otaki, H., and Otake, M. (2017). “Interactive robotic system assisting image based dialogue for the purpose of cognitive training of older adults,” in 2017 AAAI Spring Symposium Series (Stanford, CA).

Ozcana, B., Caligiorea, D., Fosch-Villarongab, E., Speratia, V., Morettac, T., and Baldassarrea, G. (2016). “Designing socially assistive robot (sar) for cognitive child-robot interaction (CCRI) with children with autism spectrum disorder-the case of “+ me,”” in 2nd International Conference on Social Robots in Therapy & Education (Barcelona).

Paletta, L., Fellner, M., Schüssler, S., Zuschnegg, J., Steiner, J., Lerch, A., et al. (2018). “Amigo: towards social robot based motivation for playful multimodal intervention in dementia,” in Proceedings of the 11th PErvasive Technologies Related to Assistive Environments Conference (Corfu), 421–427. doi: 10.1145/3197768.3203182

CrossRef Full Text | Google Scholar

Pandey, A. K., and Gelin, R. (2018). A mass-produced sociable humanoid robot: pepper: the first machine of its kind. IEEE Robot. Autom. Mag. 25, 40–48. doi: 10.1109/MRA.2018.2833157

CrossRef Full Text | Google Scholar

Paradeda, R. B., Hashemian, M., Rodrigues, R. A., and Paiva, A. (2016). “How facial expressions and small talk may influence trust in a robot,” in International Conference on Social Robotics (Kansas City, MO: Springer), 169–178. doi: 10.1007/978-3-319-47437-3_17

CrossRef Full Text | Google Scholar

Peca, A., Coeckelbergh, M., Simut, R., Costescu, C., Pintea, S., David, D., et al. (2016). Robot enhanced therapy for children with autism disorders: measuring ethical acceptability. IEEE Technol. Soc. Mag. 35, 54–66. doi: 10.1109/MTS.2016.2554701

CrossRef Full Text | Google Scholar

Peleka, G., Kargakos, A., Skartados, E., Kostavelis, I., Giakoumis, D., Sarantopoulos, I., et al. (2018). “Ramcip-a service robot for MCI patients at home,” in 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (Madrid: IEEE), 1–9. doi: 10.1109/IROS.2018.8594214

CrossRef Full Text | Google Scholar

Pennisi, P., Tonacci, A., Tartarisco, G., Billeci, L., Ruta, L., Gangemi, S., et al. (2016). Autism and social robotics: a systematic review. Autism Res. 9, 165–183. doi: 10.1002/aur.1527

PubMed Abstract | CrossRef Full Text | Google Scholar

Pereira, J., de Melo, M., Franco, N., Rodrigues, F., Coelho, A., and Fidalgo, R. (2019). “Using assistive robotics for aphasia rehabilitation,” in 2019 Latin American Robotics Symposium (LARS) 2019 Brazilian Symposium on Robotics (SBR) and 2019 Workshop on Robotics in Education (WRE) (Rio Grande: IEEE), 387–392. doi: 10.1109/LARS-SBR-WRE48964.2019.00074

CrossRef Full Text | Google Scholar

Pettinati, M. J., and Arkin, R. C. (2015). “Towards a robot computational model to preserve dignity in stigmatizing patient-caregiver relationships,” in International Conference on Social Robotics (Paris: Springer), 532–542. doi: 10.1007/978-3-319-25554-5_53

CrossRef Full Text | Google Scholar

Pierson, H. A., and Gashler, M. S. (2017). Deep learning in robotics: a review of recent research. Adv. Robot. 31, 821–835. doi: 10.1080/01691864.2017.1365009

CrossRef Full Text | Google Scholar

Pino, O., Palestra, G., Trevino, R., and De Carolis, B. (2020). The humanoid robot NAO as trainer in a memory program for elderly people with mild cognitive impairment. Int. J. Soc. Robot. 12, 21–33. doi: 10.1007/s12369-019-00533-y

CrossRef Full Text | Google Scholar

Price, C. J. (2012). A review and synthesis of the first 20 years of pet and fmri studies of heard speech, spoken language and reading. Neuroimage 62, 816–847. doi: 10.1016/j.neuroimage.2012.04.062

PubMed Abstract | CrossRef Full Text | Google Scholar

Rabbitt, S. M., Kazdin, A. E., and Hong, J. H. (2015). Acceptability of robot-assisted therapy for disruptive behavior problems in children. Archiv. Sci. Psychol. 3:101. doi: 10.1037/arc0000017

CrossRef Full Text | Google Scholar

Rahbar, F., Anzalone, S. M., Varni, G., Zibetti, E., Ivaldi, S., and Chetouani, M. (2015). “Predicting extraversion from non-verbal features during a face-to-face human-robot interaction,” in International Conference on Social Robotics (Paris: Springer), 543–553. doi: 10.1007/978-3-319-25554-5_54

CrossRef Full Text | Google Scholar

Rauchbauer, B., Nazarian, B., Bourhis, M., Ochs, M., Prévot, L., and Chaminade, T. (2019). Brain activity during reciprocal social interaction investigated using conversational robots as control condition. Philos. Trans. R. Soc. B 374:20180033. doi: 10.1098/rstb.2018.0033

PubMed Abstract | CrossRef Full Text | Google Scholar

Rehm, M., Krummheuer, A. L., Rodil, K., Nguyen, M., and Thorlacius, B. (2016). From social practices to social robots-user-driven robot development in elder care,” in International Conference on Social Robotics (Kansas City, MO: Springer), 692–701. doi: 10.1007/978-3-319-47437-3_68

CrossRef Full Text | Google Scholar

Richardson, K., Coeckelbergh, M., Wakunuma, K., Billing, E., Ziemke, T., Gomez, P., et al. (2018). Robot enhanced therapy for children with autism (dream): a social model of autism. IEEE Technol. Soc. Mag. 37, 30–39. doi: 10.1109/MTS.2018.2795096

CrossRef Full Text | Google Scholar

Riek, L., and Howard, D. (2014). “A code of ethics for the human-robot interaction profession,” in Proceedings of We Robot (Coral Gables, FL).

Google Scholar

Riek, L. D. (2016). “Robotics technology in mental health care,” in Artificial Intelligence in Behavioral and Mental Health Care (San Diego, CA: Elsevier), 185–203. doi: 10.1016/B978-0-12-420248-1.00008-8

CrossRef Full Text | Google Scholar

Riek, L. D. (2017). Healthcare robotics. Commun. ACM 60, 68–78. doi: 10.1145/3127874

CrossRef Full Text | Google Scholar

Robinette, P., Howard, A. M., and Wagner, A. R. (2015). “Timing is key for robot trust repair,” in International Conference on Social Robotics (Paris: Springer), 574–583. doi: 10.1007/978-3-319-25554-5_57

CrossRef Full Text | Google Scholar

Rudovic, O., Lee, J., Dai, M., Schuller, B., and Picard, R. W. (2018). Personalized machine learning for robot perception of affect and engagement in autism therapy. Sci. Robot. 3:eaao6760. doi: 10.1126/scirobotics.aao6760

PubMed Abstract | CrossRef Full Text | Google Scholar

Rudovic, O., Lee, J., Mascarell-Maricic, L., Schuller, B. W., and Picard, R. W. (2017). Measuring engagement in robot-assisted autism therapy: a cross-cultural study. Front. Robot. AI 4:36. doi: 10.3389/frobt.2017.00036

CrossRef Full Text | Google Scholar

Salem, M., Lakatos, G., Amirabdollahian, F., and Dautenhahn, K. (2015). “Towards safe and trustworthy social robots: ethical challenges and practical issues,” in International Conference on Social Robotics (Paris: Springer), 584–593. doi: 10.1007/978-3-319-25554-5_58

CrossRef Full Text | Google Scholar

Salichs, E., Fernández-Rodicio, E., Castillo, J. C., Castro-González, Á., Malfaz, M., and Salichs, M. Á. (2018). “A social robot assisting in cognitive stimulation therapy,” in International Conference on Practical Applications of Agents and Multi-Agent Systems (Toledo: Springer), 344–347. doi: 10.1007/978-3-319-94580-4_35

CrossRef Full Text | Google Scholar

Salichs, M. A., Encinar, I. P., Salichs, E., Castro-González, Á., and Malfaz, M. (2016). Study of scenarios and technical requirements of a social assistive robot for Alzheimer's disease patients and their caregivers. Int. J. Soc. Robot. 8, 85–102. doi: 10.1007/s12369-015-0319-6

CrossRef Full Text | Google Scholar

Salvador, M., Marsh, A. S., Gutierrez, A., and Mahoor, M. H. (2016). “Development of an aba autism intervention delivered by a humanoid robot,” in International Conference on Social Robotics (Kansas City, MO: Springer), 551–560. doi: 10.1007/978-3-319-47437-3_54

CrossRef Full Text | Google Scholar

Sandygulova, A., and O'Hare, G. M. (2015). “Children's perception of synthesized voice: robot's gender, age and accent,” in International Conference on Social Robotics (Paris: Springer), 594–602. doi: 10.1007/978-3-319-25554-5_59

CrossRef Full Text | Google Scholar

Sandygulova, A., Zhexenova, Z., Tleubayev, B., Nurakhmetova, A., Zhumabekova, D., Assylgali, I., et al. (2019). Interaction design and methodology of robot-assisted therapy for children with severe ASD and ADHD. Paladyn J. Behav. Robot. 10, 330–345. doi: 10.1515/pjbr-2019-0027

CrossRef Full Text

Santatiwongchai, S., Kaewkamnerdpong, B., Jutharee, W., and Ounjai, K. (2016). “Bliss: using robot in learning intervention to promote social skills for autism therapy,” in Proceedings of the international Convention on Rehabilitation Engineering & Assistive Technology, i-CREATe 2016 (Midview City: Singapore Therapeutic, Assistive & Rehabilitative Technologies [START] Centre), 1–4.

Google Scholar

Saxe, R., and Wexler, A. (2005). Making sense of another mind: the role of the right temporo-parietal junction. Neuropsychologia 43, 1391–1399. doi: 10.1016/j.neuropsychologia.2005.02.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Scassellati, B., Boccanfuso, L., Huang, C. M., Mademtzi, M., Qin, M., Salomons, N., et al. (2018). Improving social skills in children with asd using a long-term, in-home social robot. Sci. Robot. 3:eaat7544. doi: 10.1126/scirobotics.aat7544

PubMed Abstract | CrossRef Full Text | Google Scholar

Schüssler, S., Zuschnegg, J., Paletta, L., Fellner, M., Lodron, G., Steiner, J., et al. (2020). The effects of a humanoid socially assistive robot versus tablet training on psychosocial and physical outcomes of persons with dementia: protocol for a mixed methods study. JMIR Res. Protoc. 9:e14927. doi: 10.2196/14927

PubMed Abstract | CrossRef Full Text | Google Scholar

Shim, J., and Arkin, R. C. (2016). “Other-oriented robot deception: how can a robot's deceptive feedback help humans in HRI?” in International Conference on Social Robotics (Kansas City, MO: Springer), 222–232. doi: 10.1007/978-3-319-47437-3_22

CrossRef Full Text | Google Scholar

Shukla, J., Barreda-Ángeles, M., Oliver, J., and Puig, D. (2017). “Effectiveness of socially assistive robotics during cognitive stimulation interventions: impact on caregivers,” in 2017 26th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN) (Lisbon: IEEE), 62–67. doi: 10.1109/ROMAN.2017.8172281

CrossRef Full Text | Google Scholar

Shukla, J., Cristiano, J., Amela, D., Anguera, L., Vergés-Llahí, J., and Puig, D. (2015). “A case study of robot interaction among individuals with profound and multiple learning disabilities,” in International Conference on Social Robotics (Paris: Springer), 613–622. doi: 10.1007/978-3-319-25554-5_61

CrossRef Full Text | Google Scholar

Shukla, J., Suri, V. R., Garg, J., Verma, K., and Kansal, P. (2019). “Mapping robotic affordances with pre-requisite learning interventions for children with autism spectrum disorder,” in 2019 28th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN) (New Delhi: IEEE), 1–8. doi: 10.1109/RO-MAN46459.2019.8956247

CrossRef Full Text | Google Scholar

Siciliano, B., and Khatib, O. (2016). Springer Handbook of Robotics. Cham: Springer.

Google Scholar

Sonntag, D. (2015). “Kognit: Intelligent cognitive enhancement technology by cognitive models and mixed reality for dementia patients,” in 2015 AAAI Fall Symposium Series (Arlington, VA).

Google Scholar

Sung, H. C., Chang, S. M., Chin, M. Y., and Lee, W. L. (2015). Robot-assisted therapy for improving social interactions and activity participation among institutionalized older adults: a pilot study. Asia Pac. Psychiatry 7, 1–6. doi: 10.1111/appy.12131

PubMed Abstract | CrossRef Full Text | Google Scholar

Sutton, R. S., and Barto, A. G. (2018). Reinforcement Learning: An Introduction. Cambridge, MA: MIT Press.

Google Scholar

Taheri, A., Alemi, M., Meghdari, A., Pouretemad, H., Basiri, N. M., and Poorgoldooz, P. (2015a). “Impact of humanoid social robots on treatment of a pair of iranian autistic twins,” in International Conference on Social Robotics (Paris: Springer), 623–632. doi: 10.1007/978-3-319-25554-5_62

CrossRef Full Text | Google Scholar

Taheri, A., Alemi, M., Meghdari, A., Pouretemad, H., and Holderread, S. (2015b). Clinical application of humanoid robots in playing imitation games for autistic children in Iran. Proc. Soc. Behav. Sci. 176, 898–906. doi: 10.1016/j.sbspro.2015.01.556

CrossRef Full Text | Google Scholar

Taheri, A., Meghdari, A., Alemi, M., and Pouretemad, H. (2018). Clinical interventions of social humanoid robots in the treatment of a set of high-and low-functioning autistic iranian twins. Sci. Iranica 25, 1197–1214. doi: 10.24200/SCI.2017.4337

CrossRef Full Text | Google Scholar

Taheri, A., Meghdari, A., Alemi, M., and Pouretemad, H. (2019). Teaching music to children with autism: a social robotics challenge. Sci. Iranica 26, 40–58. doi: 10.24200/sci.2017.4608

CrossRef Full Text | Google Scholar

Taheri, A., Meghdari, A., Alemi, M., Pouretemad, H., Poorgoldooz, P., and Roohbakhsh, M. (2016). “Social robots and teaching music to autistic children: myth or reality?” in International Conference on Social Robotics (Kansas City, MO: Springer), 541–550. doi: 10.1007/978-3-319-47437-3_53

CrossRef Full Text | Google Scholar

Taheri, A., Meghdari, A., and Mahoor, M. H. (2020). A close look at the imitation performance of children with autism and typically developing children using a robotic system. Int. J. Soc. Robot. doi: 10.1007/s12369-020-00704-2. [Epub ahead of print].

CrossRef Full Text | Google Scholar

Taranović, A., Jevtić, A., Hernández-Farigola, J., Tantinyà, N., Abdelnour, C., and Torras, C. (2018a). “Resource-based modality selection in robot-assisted cognitive training,” in Companion of the 2018 ACM/IEEE International Conference on Human-Robot Interaction (Chicago, IL), 255–256. doi: 10.1145/3173386.3176961

CrossRef Full Text | Google Scholar

Taranović, A., Jevtić, A., and Torras, C. (2018b). “Adaptable multimodal interaction framework for robot-assisted cognitive training,” in Companion of the 2018 ACM/IEEE International Conference on Human-Robot Interaction (Chicago, IL), 327–328. doi: 10.1145/3173386.3176911

CrossRef Full Text | Google Scholar

Taranović, A., Jevtić, A., and Torras, C. (2018c). “Adaptive modality selection algorithm in robot-assisted cognitive training,” in 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (Madrid: IEEE), 4456–4461. doi: 10.1109/IROS.2018.8593730

CrossRef Full Text | Google Scholar

Tariq, S., Baber, S., Ashfaq, A., Ayaz, Y., Naveed, M., and Mohsin, S. (2016). “Interactive therapy approach through collaborative physical play between a socially assistive humanoid robot and children with autism spectrum disorder,” in International Conference on Social Robotics (Kansas City, MO: Springer), 561–570. doi: 10.1007/978-3-319-47437-3_55

CrossRef Full Text | Google Scholar

Tleubayev, B., Zhexenova, Z., Zhakenova, A., and Sandygulova, A. (2019). “Robot-assisted therapy for children with adhd and asd: a pilot study,” in Proceedings of the 2019 2nd International Conference on Service Robotics Technologies (Beijing), 58–62. doi: 10.1145/3325693.3325703

CrossRef Full Text | Google Scholar

Tokunaga, S., Seaborn, K., Tamura, K., and Otake-Matsuura, M. (2019). “Cognitive training for older adults with a dialogue-based, robot-facilitated storytelling system,” in International Conference on Interactive Digital Storytelling (Little Cottonwood Canyon, UT: Springer), 405–409. doi: 10.1007/978-3-030-33894-7_43

CrossRef Full Text | Google Scholar

Trovato, G., Galeazzi, J., Torta, E., Ham, J. R., and Cuijpers, R. H. (2015). “Study on adaptation of robot communication strategies in changing situations,” in International Conference on Social Robotics (Paris: Springer), 654–663. doi: 10.1007/978-3-319-25554-5_65

CrossRef Full Text | Google Scholar

Tsardoulias, E. G., Kintsakis, A. M., Panayiotou, K., Thallas, A. G., Reppou, S. E., Karagiannis, G. G., et al. (2017). Towards an integrated robotics architecture for social inclusion-the rapp paradigm. Cogn. Syst. Res. 43, 157–173. doi: 10.1016/j.cogsys.2016.08.004

CrossRef Full Text | Google Scholar

Tsiakas, K., Abujelala, M., and Makedon, F. (2018). Task engagement as personalization feedback for socially-assistive robots and cognitive training. Technologies 6:49. doi: 10.3390/technologies6020049

CrossRef Full Text | Google Scholar

Tsiakas, K., Dagioglou, M., Karkaletsis, V., and Makedon, F. (2016). “Adaptive robot assisted therapy using interactive reinforcement learning,” in International Conference on Social Robotics (Kansas City, MO: Springer), 11–21. doi: 10.1007/978-3-319-47437-3_2

CrossRef Full Text | Google Scholar

Tung, F. W. (2016). Child perception of humanoid robot appearance and behavior. Int. J. Hum. Comput. Interact. 32, 493–502. doi: 10.1080/10447318.2016.1172808

CrossRef Full Text | Google Scholar

University of Minnesota Libraries (2021). How Does Google Scholar Work? Minneapolis, MN: University of Minnesota Libraries.

Valentí Soler, M., Agüera-Ortiz, L., Olazarán Rodríguez, J., Mendoza Rebolledo, C., Pérez Muñoz, A., Rodríguez Pérez, I., et al. (2015). Social robots in advanced dementia. Front. Aging Neurosci. 7:133. doi: 10.3389/fnagi.2015.00133

PubMed Abstract | CrossRef Full Text | Google Scholar

van den Heuvel, R. J., Lexis, M. A., and de Witte, L. P. (2017). Robot zora in rehabilitation and special education for children with severe physical disabilities: a pilot study. Int. J. Rehabil. Res. 40:353. doi: 10.1097/MRR.0000000000000248

PubMed Abstract | CrossRef Full Text | Google Scholar

Vaufreydaz, D., Johal, W., and Combe, C. (2016). Starting engagement detection towards a companion robot using multimodal features. Robot. Auton. Syst. 75, 4–16. doi: 10.1016/j.robot.2015.01.004

CrossRef Full Text | Google Scholar

Veerbeek, J. M., Langbroek-Amersfoort, A. C., Van Wegen, E. E., Meskers, C. G., and Kwakkel, G. (2017). Effects of robot-assisted therapy for the upper limb after stroke: a systematic review and meta-analysis. Neurorehabil. Neural Repair 31, 107–121. doi: 10.1177/1545968316666957

CrossRef Full Text | Google Scholar

Villaronga, E. F. (2016). “What do roboticists need to know about the future of robot law,” in New Friends Conference Proceedings, New Friends: 2nd International Conference on Social Robots in Therapy and Education (Barcelona).

Google Scholar

Wan, S., Gu, Z., and Ni, Q. (2020). Cognitive computing and wireless communications on the edge for healthcare service robots. Comput. Commun. 149, 99–106. doi: 10.1016/j.comcom.2019.10.012

CrossRef Full Text | Google Scholar

Weiss, W., Fuhrmann, F., Zeiner, H., and Unterberger, R. (2017). “Towards an architecture for collaborative human robot interaction in physiotherapeutic applications,” in Proceedings of the Companion of the 2017 ACM/IEEE International Conference on Human-Robot Interaction (Vienna), 319–320. doi: 10.1145/3029798.3038393

CrossRef Full Text | Google Scholar

Whelan, S., Murphy, K., Barrett, E., Krusche, C., Santorelli, A., and Casey, D. (2018). Factors affecting the acceptability of social robots by older adults including people with dementia or cognitive impairment: a literature review. Int. J. Soc. Robot. 10, 643–668. doi: 10.1007/s12369-018-0471-x

CrossRef Full Text | Google Scholar

WHO (2011). World Report on Disability 2011. Geneva: World Health Organization.

Google Scholar

Wiltshire, T. J., Warta, S. F., Barber, D., and Fiore, S. M. (2017). Enabling robotic social intelligence by engineering human social-cognitive mechanisms. Cogn. Syst. Res. 43, 190–207. doi: 10.1016/j.cogsys.2016.09.005

CrossRef Full Text | Google Scholar

Wong, H., and Zhong, Z. (2016). “Assessment of robot training for social cognitive learning,” in 2016 16th International Conference on Control, Automation and Systems (ICCAS) (Gyeongju: IEEE), 893–898. doi: 10.1109/ICCAS.2016.7832420

CrossRef Full Text | Google Scholar

Woo, J., Botzheim, J., and Kubota, N. (2017). System integration for cognitive model of a robot partner. Intell. Autom. Soft Comput. 4, 1–14. doi: 10.1080/10798587.2017.1364919

CrossRef Full Text | Google Scholar

Wood, L. J., Zaraki, A., Walters, M. L., Novanda, O., Robins, B., and Dautenhahn, K. (2017). “The iterative development of the humanoid robot Kaspar: an assistive robot for children with autism,” in International Conference on Social Robotics (Tsukuba: Springer), 53–63. doi: 10.1007/978-3-319-70022-9_6

CrossRef Full Text | Google Scholar

Yokota, K., Ooi, S., and Sano, M. (2019). “A method of generating a dialogue pattern to induce awareness based on a reflection support agent,” in International Conference on Human-Computer Interaction (Orlando, FL: Springer), 421–438. doi: 10.1007/978-3-030-22219-2_32

CrossRef Full Text | Google Scholar

Yu, R., Hui, E., Lee, J., Poon, D., Ng, A., Sit, K., et al. (2015). Use of a therapeutic, socially assistive pet robot (Paro) in improving mood and stimulating social interaction and communication for people with dementia: study protocol for a randomized controlled trial. JMIR Res. Protoc. 4:e45. doi: 10.2196/resprot.4189

PubMed Abstract | CrossRef Full Text | Google Scholar

Yun, S. S., Kim, H., Choi, J., and Park, S. K. (2016). A robot-assisted behavioral intervention system for children with autism spectrum disorders. Robot. Auton. Syst. 76, 58–67. doi: 10.1016/j.robot.2015.11.004

CrossRef Full Text | Google Scholar

Zheng, Z., Nie, G., Swanson, A., Weitlauf, A., Warren, Z., and Sarkar, N. (2016). “Longitudinal impact of autonomous robot-mediated joint attention intervention for young children with ASD,” in International Conference on Social Robotics (Kansas City, MO: Springer), 581–590. doi: 10.1007/978-3-319-47437-3_57

CrossRef Full Text | Google Scholar

Zheng, Z., Young, E. M., Swanson, A. R., Weitlauf, A. S., Warren, Z. E., and Sarkar, N. (2015). Robot-mediated imitation skill training for children with autism. IEEE Trans. Neural Syst. Rehabil. Eng. 24, 682–691. doi: 10.1109/TNSRE.2015.2475724

PubMed Abstract | CrossRef Full Text | Google Scholar

Zsiga, K., Tóth, A., Pilissy, T., Péter, O., Dénes, Z., and Fazekas, G. (2018). Evaluation of a companion robot based on field tests with single older adults in their homes. Assist. Technol. 30, 259–266. doi: 10.1080/10400435.2017.1322158

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: rehabilitation robotics, human-robot interaction, robot-assisted cognitive training, socially assistive robotics, multimodal interaction, cognitive disability

Citation: Yuan F, Klavon E, Liu Z, Lopez RP and Zhao X (2021) A Systematic Review of Robotic Rehabilitation for Cognitive Training. Front. Robot. AI 8:605715. doi: 10.3389/frobt.2021.605715

Received: 13 September 2020; Accepted: 06 April 2021;
Published: 11 May 2021.

Edited by:

Raquel Ros, Universitat Ramon Llull, Spain

Reviewed by:

Soheil Keshmiri, Advanced Telecommunications Research Institute International (ATR), Japan
Guillem Alenyà, Consejo Superior de Investigaciones Científicas (CSIC), Spain
Antonio Andriella, Instituto de Robótica e Informática Industrial (IRII), Spain, in collaboration with reviewer GA

Copyright © 2021 Yuan, Klavon, Liu, Lopez and Zhao. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Xiaopeng Zhao, xzhao9@utk.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.